var/home/core/zuul-output/0000755000175000017500000000000015116003661014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116016302015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005232616215116016270017703 0ustar rootrootDec 09 11:31:13 crc systemd[1]: Starting Kubernetes Kubelet... Dec 09 11:31:13 crc restorecon[4755]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:13 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 11:31:14 crc restorecon[4755]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 09 11:31:15 crc kubenswrapper[4770]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.167037 4770 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171229 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171251 4770 feature_gate.go:330] unrecognized feature gate: Example Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171258 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171264 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171269 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171275 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171280 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171287 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171294 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171300 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171305 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171311 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171316 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171329 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171333 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171338 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171342 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171346 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171351 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171355 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171360 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171365 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171369 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171374 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171380 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171385 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171390 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171395 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171399 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171404 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171408 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171413 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171416 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171421 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171425 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171431 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171436 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171441 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171445 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171450 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171454 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171459 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171463 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171467 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171474 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171478 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171483 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171487 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171492 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171497 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171502 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171506 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171510 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171514 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171519 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171523 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171527 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171531 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171535 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171541 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171547 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171552 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171557 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171562 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171566 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171572 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171576 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171581 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171586 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171590 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.171594 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172164 4770 flags.go:64] FLAG: --address="0.0.0.0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172180 4770 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172189 4770 flags.go:64] FLAG: --anonymous-auth="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172195 4770 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172202 4770 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172209 4770 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172218 4770 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172224 4770 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172229 4770 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172234 4770 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172239 4770 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172245 4770 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172250 4770 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172255 4770 flags.go:64] FLAG: --cgroup-root="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172261 4770 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172266 4770 flags.go:64] FLAG: --client-ca-file="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172271 4770 flags.go:64] FLAG: --cloud-config="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172275 4770 flags.go:64] FLAG: --cloud-provider="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172279 4770 flags.go:64] FLAG: --cluster-dns="[]" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172286 4770 flags.go:64] FLAG: --cluster-domain="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172291 4770 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172296 4770 flags.go:64] FLAG: --config-dir="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172301 4770 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172306 4770 flags.go:64] FLAG: --container-log-max-files="5" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172313 4770 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172318 4770 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172324 4770 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172329 4770 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172334 4770 flags.go:64] FLAG: --contention-profiling="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172339 4770 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172343 4770 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172349 4770 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172354 4770 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172361 4770 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172365 4770 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172370 4770 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172375 4770 flags.go:64] FLAG: --enable-load-reader="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172381 4770 flags.go:64] FLAG: --enable-server="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172386 4770 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172392 4770 flags.go:64] FLAG: --event-burst="100" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172398 4770 flags.go:64] FLAG: --event-qps="50" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172403 4770 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172408 4770 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172413 4770 flags.go:64] FLAG: --eviction-hard="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172420 4770 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172425 4770 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172430 4770 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172437 4770 flags.go:64] FLAG: --eviction-soft="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172442 4770 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172447 4770 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172454 4770 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172459 4770 flags.go:64] FLAG: --experimental-mounter-path="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172464 4770 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172468 4770 flags.go:64] FLAG: --fail-swap-on="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172475 4770 flags.go:64] FLAG: --feature-gates="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172481 4770 flags.go:64] FLAG: --file-check-frequency="20s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172486 4770 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172491 4770 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172497 4770 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172502 4770 flags.go:64] FLAG: --healthz-port="10248" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172507 4770 flags.go:64] FLAG: --help="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172512 4770 flags.go:64] FLAG: --hostname-override="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172516 4770 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172521 4770 flags.go:64] FLAG: --http-check-frequency="20s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172526 4770 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172530 4770 flags.go:64] FLAG: --image-credential-provider-config="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172535 4770 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172539 4770 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172544 4770 flags.go:64] FLAG: --image-service-endpoint="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172551 4770 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172556 4770 flags.go:64] FLAG: --kube-api-burst="100" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172562 4770 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172569 4770 flags.go:64] FLAG: --kube-api-qps="50" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172574 4770 flags.go:64] FLAG: --kube-reserved="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172579 4770 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172584 4770 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172590 4770 flags.go:64] FLAG: --kubelet-cgroups="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172594 4770 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172600 4770 flags.go:64] FLAG: --lock-file="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172605 4770 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172610 4770 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172616 4770 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172625 4770 flags.go:64] FLAG: --log-json-split-stream="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172630 4770 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172637 4770 flags.go:64] FLAG: --log-text-split-stream="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172642 4770 flags.go:64] FLAG: --logging-format="text" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172658 4770 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172664 4770 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172669 4770 flags.go:64] FLAG: --manifest-url="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172675 4770 flags.go:64] FLAG: --manifest-url-header="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172683 4770 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172688 4770 flags.go:64] FLAG: --max-open-files="1000000" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172694 4770 flags.go:64] FLAG: --max-pods="110" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172699 4770 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172704 4770 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172709 4770 flags.go:64] FLAG: --memory-manager-policy="None" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172714 4770 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172719 4770 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172724 4770 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172733 4770 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172747 4770 flags.go:64] FLAG: --node-status-max-images="50" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172753 4770 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172758 4770 flags.go:64] FLAG: --oom-score-adj="-999" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172762 4770 flags.go:64] FLAG: --pod-cidr="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172767 4770 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172777 4770 flags.go:64] FLAG: --pod-manifest-path="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172782 4770 flags.go:64] FLAG: --pod-max-pids="-1" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172787 4770 flags.go:64] FLAG: --pods-per-core="0" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172791 4770 flags.go:64] FLAG: --port="10250" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172796 4770 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172801 4770 flags.go:64] FLAG: --provider-id="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172806 4770 flags.go:64] FLAG: --qos-reserved="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172811 4770 flags.go:64] FLAG: --read-only-port="10255" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172816 4770 flags.go:64] FLAG: --register-node="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172821 4770 flags.go:64] FLAG: --register-schedulable="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172826 4770 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172842 4770 flags.go:64] FLAG: --registry-burst="10" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172847 4770 flags.go:64] FLAG: --registry-qps="5" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172853 4770 flags.go:64] FLAG: --reserved-cpus="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172857 4770 flags.go:64] FLAG: --reserved-memory="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172864 4770 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172869 4770 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172875 4770 flags.go:64] FLAG: --rotate-certificates="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172880 4770 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172885 4770 flags.go:64] FLAG: --runonce="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172891 4770 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172900 4770 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172921 4770 flags.go:64] FLAG: --seccomp-default="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172927 4770 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172932 4770 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172937 4770 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172943 4770 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172951 4770 flags.go:64] FLAG: --storage-driver-password="root" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172956 4770 flags.go:64] FLAG: --storage-driver-secure="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172961 4770 flags.go:64] FLAG: --storage-driver-table="stats" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172966 4770 flags.go:64] FLAG: --storage-driver-user="root" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172971 4770 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172976 4770 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172981 4770 flags.go:64] FLAG: --system-cgroups="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172985 4770 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172994 4770 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.172999 4770 flags.go:64] FLAG: --tls-cert-file="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173003 4770 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173010 4770 flags.go:64] FLAG: --tls-min-version="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173015 4770 flags.go:64] FLAG: --tls-private-key-file="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173020 4770 flags.go:64] FLAG: --topology-manager-policy="none" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173024 4770 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173030 4770 flags.go:64] FLAG: --topology-manager-scope="container" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173034 4770 flags.go:64] FLAG: --v="2" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173041 4770 flags.go:64] FLAG: --version="false" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173047 4770 flags.go:64] FLAG: --vmodule="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173053 4770 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173058 4770 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173188 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173194 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173199 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173203 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173208 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173213 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173219 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173224 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173229 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173233 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173237 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173244 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173249 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173253 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173258 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173262 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173266 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173270 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173274 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173278 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173282 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173286 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173290 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173294 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173298 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173302 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173307 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173311 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173315 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173320 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173324 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173328 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173332 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173336 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173340 4770 feature_gate.go:330] unrecognized feature gate: Example Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173345 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173350 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173354 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173359 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173363 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173369 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173374 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173379 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173384 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173390 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173397 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173402 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173407 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173412 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173417 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173421 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173427 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173433 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173437 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173442 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173447 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173451 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173455 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173460 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173466 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173472 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173478 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173482 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173487 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173491 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173495 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173500 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173504 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173508 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173513 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.173517 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.173533 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.184590 4770 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.184647 4770 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184749 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184759 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184765 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184772 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184777 4770 feature_gate.go:330] unrecognized feature gate: Example Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184783 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184788 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184794 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184800 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184805 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184810 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184818 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184829 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184836 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184843 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184849 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184856 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184862 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184869 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184875 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184881 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184887 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184893 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184931 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184940 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184947 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184955 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184962 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184970 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184977 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184985 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.184992 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185001 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185010 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185017 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185026 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185035 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185044 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185054 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185062 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185070 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185078 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185085 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185092 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185099 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185106 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185112 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185119 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185126 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185133 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185140 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185147 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185157 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185167 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185174 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185182 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185189 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185195 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185202 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185208 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185216 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185222 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185229 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185236 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185243 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185249 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185256 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185265 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185273 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185280 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185287 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.185301 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185469 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185481 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185487 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185493 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185498 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185505 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185512 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185518 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185524 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185529 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185535 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185541 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185546 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185551 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185558 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185564 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185569 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185574 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185579 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185585 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185590 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185595 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185600 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185605 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185611 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185616 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185621 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185626 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185631 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185638 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185643 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185648 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185655 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185660 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185665 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185671 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185676 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185683 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185689 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185696 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185702 4770 feature_gate.go:330] unrecognized feature gate: Example Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185707 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185713 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185718 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185723 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185729 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185734 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185739 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185746 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185751 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185757 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185762 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185767 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185773 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185778 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185783 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185788 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185793 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185799 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185804 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185809 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185814 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185821 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185827 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185833 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185838 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185844 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185850 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185855 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185860 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.185865 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.185874 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.186124 4770 server.go:940] "Client rotation is on, will bootstrap in background" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.189639 4770 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.189747 4770 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.190560 4770 server.go:997] "Starting client certificate rotation" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.190609 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.191221 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-04 07:24:38.540787649 +0000 UTC Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.191349 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.208238 4770 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.211111 4770 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.212577 4770 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.221414 4770 log.go:25] "Validated CRI v1 runtime API" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.241274 4770 log.go:25] "Validated CRI v1 image API" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.243391 4770 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.246288 4770 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-09-11-26-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.246330 4770 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.269975 4770 manager.go:217] Machine: {Timestamp:2025-12-09 11:31:15.261690618 +0000 UTC m=+0.502449147 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c341d7e1-5989-4099-a8b0-1ad7487a3271 BootID:f08a8155-c275-4917-b646-2fe0ff409bb7 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:4e:b7:cd Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:4e:b7:cd Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d7:ba:26 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:73:36:dd Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:2c:98:5b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5c:3b:a5 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:58:b5:44 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0e:ad:5a:3c:3b:3d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f6:da:00:ef:4c:b0 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.270309 4770 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.270589 4770 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.270980 4770 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.271184 4770 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.271225 4770 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.271495 4770 topology_manager.go:138] "Creating topology manager with none policy" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.271510 4770 container_manager_linux.go:303] "Creating device plugin manager" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.271693 4770 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.272260 4770 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.272646 4770 state_mem.go:36] "Initialized new in-memory state store" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.272753 4770 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.273413 4770 kubelet.go:418] "Attempting to sync node with API server" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.273431 4770 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.273478 4770 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.273495 4770 kubelet.go:324] "Adding apiserver pod source" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.273508 4770 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.275242 4770 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.275689 4770 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.276166 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.276270 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.276178 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.276349 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.276866 4770 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277495 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277519 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277529 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277549 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277563 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277574 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277582 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277604 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277614 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277621 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277634 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.277641 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.278983 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.279534 4770 server.go:1280] "Started kubelet" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.280050 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.280302 4770 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.280303 4770 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.280880 4770 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 09 11:31:15 crc systemd[1]: Started Kubernetes Kubelet. Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.281562 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.281594 4770 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.281624 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:08:26.529123653 +0000 UTC Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.282073 4770 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.282115 4770 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.282121 4770 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.282233 4770 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.282387 4770 server.go:460] "Adding debug handlers to kubelet server" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.282838 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f88b1972e096a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 11:31:15.279489386 +0000 UTC m=+0.520247905,LastTimestamp:2025-12-09 11:31:15.279489386 +0000 UTC m=+0.520247905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.285180 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.285214 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.285275 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.292863 4770 factory.go:153] Registering CRI-O factory Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.292927 4770 factory.go:221] Registration of the crio container factory successfully Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.294749 4770 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.294804 4770 factory.go:55] Registering systemd factory Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.294820 4770 factory.go:221] Registration of the systemd container factory successfully Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.294868 4770 factory.go:103] Registering Raw factory Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.294958 4770 manager.go:1196] Started watching for new ooms in manager Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.297098 4770 manager.go:319] Starting recovery of all containers Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.306526 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.306736 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.306821 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.306895 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.306976 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307048 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307105 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307182 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307244 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307312 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307374 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307432 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307490 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307580 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307652 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307721 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307785 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307852 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.307960 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308025 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308097 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308161 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308218 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308278 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308334 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308417 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308538 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308604 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308663 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308718 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308807 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308875 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.308952 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309013 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309069 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309135 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309198 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309258 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309318 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309397 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309462 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309520 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309592 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309655 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309712 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309772 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309830 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.309952 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310038 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310100 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310160 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310215 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310282 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310345 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310401 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310457 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310513 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310569 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310695 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310755 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310813 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310895 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.310977 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311047 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311108 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311166 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311224 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311287 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311353 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311414 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311470 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311526 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311614 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311685 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311754 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311811 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311874 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.311962 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.312027 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.312128 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.312192 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.312249 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.312321 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.313170 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.313551 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.313639 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.313718 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.313791 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315316 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315424 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315515 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315633 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315719 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315793 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315878 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.315984 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.316943 4770 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318053 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318117 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318143 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318157 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318171 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318185 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318217 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318229 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318257 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318274 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318298 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318329 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318358 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318375 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318390 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318404 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318423 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318436 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318448 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318460 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318474 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318484 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318494 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318504 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318517 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318538 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318552 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318570 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318590 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318610 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318626 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318640 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318656 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318668 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318684 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318698 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318709 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318720 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318736 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318759 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318793 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318808 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318820 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318831 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318841 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318853 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318865 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318878 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318894 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318959 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318982 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.318997 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319016 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319038 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319051 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319094 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319104 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319119 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319135 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319146 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319157 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319167 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319178 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319189 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319201 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319215 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319227 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319237 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319254 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319265 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319278 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319318 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319334 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319361 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319383 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319393 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319406 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319428 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319455 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319468 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319480 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319491 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319505 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319527 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319540 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319555 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319677 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319713 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319730 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319743 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319753 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319767 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319782 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319793 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319814 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319827 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319838 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319854 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319871 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319882 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319892 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319931 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319949 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319965 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319979 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.319994 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320009 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320025 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320041 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320061 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320073 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320083 4770 reconstruct.go:97] "Volume reconstruction finished" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.320092 4770 reconciler.go:26] "Reconciler: start to sync state" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.323049 4770 manager.go:324] Recovery completed Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.334195 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.336338 4770 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.337727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.337773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.337803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338634 4770 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338678 4770 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338861 4770 kubelet.go:2335] "Starting kubelet main sync loop" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338920 4770 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338933 4770 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.338960 4770 state_mem.go:36] "Initialized new in-memory state store" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.338929 4770 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.339516 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.339587 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.350308 4770 policy_none.go:49] "None policy: Start" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.351628 4770 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.351676 4770 state_mem.go:35] "Initializing new in-memory state store" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.382795 4770 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.404271 4770 manager.go:334] "Starting Device Plugin manager" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.404332 4770 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.404349 4770 server.go:79] "Starting device plugin registration server" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.404861 4770 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.404880 4770 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.405111 4770 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.405211 4770 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.405226 4770 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.410769 4770 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.439384 4770 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.439516 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.442468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.442524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.442591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.443569 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.443653 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.443710 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.444676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.444708 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.444721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.444863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.444966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.445034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.445229 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.445382 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.445413 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446639 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.446787 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447044 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447117 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447703 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.447994 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.448078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.448096 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.448148 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.448242 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.448265 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449053 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449098 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449117 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449268 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449288 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449846 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449878 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.449891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.485979 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.505729 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.507343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.507389 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.507400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.507426 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.508118 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522816 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522854 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522873 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522892 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522946 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.522996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523072 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523149 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523228 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523304 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523330 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.523352 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624684 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624708 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624746 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624765 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624780 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624812 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624819 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624868 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624849 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624921 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624893 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624827 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624941 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.624990 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625041 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625102 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625125 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625150 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625175 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625204 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625189 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625217 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.625240 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.708890 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.710167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.710200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.710212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.710236 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.710834 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.771107 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.784092 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.791224 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.795584 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-56347407758e3dbf2ef66abee21dc3500a27f3d840c4b54d4b34db4bf9911472 WatchSource:0}: Error finding container 56347407758e3dbf2ef66abee21dc3500a27f3d840c4b54d4b34db4bf9911472: Status 404 returned error can't find the container with id 56347407758e3dbf2ef66abee21dc3500a27f3d840c4b54d4b34db4bf9911472 Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.803926 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-a83334528eab1d2f4843688b183946eb4e6558833987fe9e44fbadb4a41b6971 WatchSource:0}: Error finding container a83334528eab1d2f4843688b183946eb4e6558833987fe9e44fbadb4a41b6971: Status 404 returned error can't find the container with id a83334528eab1d2f4843688b183946eb4e6558833987fe9e44fbadb4a41b6971 Dec 09 11:31:15 crc kubenswrapper[4770]: W1209 11:31:15.811823 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c0aa27fd5bf6212d85abd58cac8bfd1ea7c4e7409209ca5e263ec47f28a7ad26 WatchSource:0}: Error finding container c0aa27fd5bf6212d85abd58cac8bfd1ea7c4e7409209ca5e263ec47f28a7ad26: Status 404 returned error can't find the container with id c0aa27fd5bf6212d85abd58cac8bfd1ea7c4e7409209ca5e263ec47f28a7ad26 Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.813088 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: I1209 11:31:15.816940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:15 crc kubenswrapper[4770]: E1209 11:31:15.887180 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.111636 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.113779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.113828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.113840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.113867 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.114447 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 09 11:31:16 crc kubenswrapper[4770]: W1209 11:31:16.260402 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.260475 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.281358 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.282337 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 23:14:18.563059382 +0000 UTC Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.344857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.345009 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c0aa27fd5bf6212d85abd58cac8bfd1ea7c4e7409209ca5e263ec47f28a7ad26"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.345188 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.346315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.346355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.346367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.347085 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58" exitCode=0 Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.347156 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.347180 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a83334528eab1d2f4843688b183946eb4e6558833987fe9e44fbadb4a41b6971"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.347269 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.348002 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.348030 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.348039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.349458 4770 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6" exitCode=0 Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.349514 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.349538 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"56347407758e3dbf2ef66abee21dc3500a27f3d840c4b54d4b34db4bf9911472"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.349657 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350668 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350715 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b68df18bd99e9ff924def6e4a005e9f449259f8605f0728d3d32439d07522880"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.350832 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.352255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.352299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.352312 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.352462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3"} Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.352494 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6e3e5fb9c4e55bf0be2bcae9a706b3fa3b42c058f08e361ca54b9af0335c4a12"} Dec 09 11:31:16 crc kubenswrapper[4770]: W1209 11:31:16.587066 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.587153 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:16 crc kubenswrapper[4770]: W1209 11:31:16.607227 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.607292 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:16 crc kubenswrapper[4770]: W1209 11:31:16.688463 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.688542 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.688499 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.736144 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f88b1972e096a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 11:31:15.279489386 +0000 UTC m=+0.520247905,LastTimestamp:2025-12-09 11:31:15.279489386 +0000 UTC m=+0.520247905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.915277 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.916501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.916533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.916566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:16 crc kubenswrapper[4770]: I1209 11:31:16.916590 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:16 crc kubenswrapper[4770]: E1209 11:31:16.916968 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.281344 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.282667 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 15:38:38.79123935 +0000 UTC Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.355308 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.356149 4770 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea" exitCode=0 Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.356227 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea"} Dec 09 11:31:17 crc kubenswrapper[4770]: E1209 11:31:17.356235 4770 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.356345 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.357140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.357167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.357176 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.362736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.362933 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.362963 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.363497 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.364485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.364509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.364519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.365010 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a" exitCode=0 Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.365096 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.365199 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366031 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366049 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366649 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b" exitCode=0 Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.366812 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.367354 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.367377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.367394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.367601 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.368505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.368533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.368544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.368822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671"} Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.368943 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.369633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.369659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:17 crc kubenswrapper[4770]: I1209 11:31:17.369669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: W1209 11:31:18.008495 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 09 11:31:18 crc kubenswrapper[4770]: E1209 11:31:18.008920 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.283557 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 20:47:16.394626518 +0000 UTC Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.283611 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 273h15m58.111018084s for next certificate rotation Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.374257 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d" exitCode=0 Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.374339 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.374478 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.375268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.375315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.375328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.377806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.377845 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.377861 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.377963 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.378873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.378920 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.378974 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380697 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380740 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380756 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380776 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380800 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8"} Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.380980 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.384352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.384734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.385470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.385496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.386151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.386202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.517260 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.518262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.518291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.518300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:18 crc kubenswrapper[4770]: I1209 11:31:18.518321 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386577 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e"} Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386624 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386672 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc"} Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8"} Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.386754 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85"} Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.387542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.387579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.387594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.988960 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.989216 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.989368 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.990659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.990701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:19 crc kubenswrapper[4770]: I1209 11:31:19.990713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.394293 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7"} Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.394403 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.394427 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.395768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:20 crc kubenswrapper[4770]: I1209 11:31:20.976206 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.397253 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.397555 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.398815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:21 crc kubenswrapper[4770]: I1209 11:31:21.504716 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.435182 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.435367 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.436646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.436686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.436696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:22 crc kubenswrapper[4770]: I1209 11:31:22.461999 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.239751 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.246175 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.394019 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.394216 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.395279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.395311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.395324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.401127 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.401953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.401971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.401979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.927167 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.927454 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.929139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.929268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:23 crc kubenswrapper[4770]: I1209 11:31:23.929356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:24 crc kubenswrapper[4770]: I1209 11:31:24.404238 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:24 crc kubenswrapper[4770]: I1209 11:31:24.405370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:24 crc kubenswrapper[4770]: I1209 11:31:24.405426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:24 crc kubenswrapper[4770]: I1209 11:31:24.405436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:25 crc kubenswrapper[4770]: E1209 11:31:25.410874 4770 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.435249 4770 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.435351 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.704035 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.704236 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.705411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.705457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:25 crc kubenswrapper[4770]: I1209 11:31:25.705486 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:26 crc kubenswrapper[4770]: I1209 11:31:26.887702 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 09 11:31:26 crc kubenswrapper[4770]: I1209 11:31:26.887923 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:26 crc kubenswrapper[4770]: I1209 11:31:26.889296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:26 crc kubenswrapper[4770]: I1209 11:31:26.889345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:26 crc kubenswrapper[4770]: I1209 11:31:26.889360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:27 crc kubenswrapper[4770]: I1209 11:31:27.787875 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]log ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]etcd ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-filter ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-apiextensions-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/start-apiextensions-controllers failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/crd-informer-synced failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-system-namespaces-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/start-service-ip-repair-controllers failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/priority-and-fairness-config-producer failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/bootstrap-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-aggregator-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/apiservice-registration-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/apiservice-discovery-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]autoregister-completion failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapi-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: livez check failed Dec 09 11:31:27 crc kubenswrapper[4770]: I1209 11:31:27.787994 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:31:27 crc kubenswrapper[4770]: I1209 11:31:27.794763 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]log ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]etcd ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-filter ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-apiextensions-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/start-apiextensions-controllers failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/crd-informer-synced failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-system-namespaces-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/start-service-ip-repair-controllers failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/priority-and-fairness-config-producer failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/bootstrap-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/start-kube-aggregator-informers ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/apiservice-registration-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 09 11:31:27 crc kubenswrapper[4770]: [-]poststarthook/apiservice-discovery-controller failed: reason withheld Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]autoregister-completion ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapi-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 09 11:31:27 crc kubenswrapper[4770]: livez check failed Dec 09 11:31:27 crc kubenswrapper[4770]: I1209 11:31:27.794835 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.994247 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.994465 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.995199 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.995266 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.995681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.995721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:29 crc kubenswrapper[4770]: I1209 11:31:29.995734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.000610 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.417678 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.418243 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.418292 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.418469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.418501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:30 crc kubenswrapper[4770]: I1209 11:31:30.418512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:31 crc kubenswrapper[4770]: I1209 11:31:31.381859 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 11:31:31 crc kubenswrapper[4770]: I1209 11:31:31.381994 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.466284 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.466428 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.469466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.470334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.470351 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.768963 4770 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 11:31:32 crc kubenswrapper[4770]: E1209 11:31:32.769410 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.769924 4770 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.771081 4770 trace.go:236] Trace[1876249449]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 11:31:19.109) (total time: 13661ms): Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[1876249449]: ---"Objects listed" error: 13661ms (11:31:32.771) Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[1876249449]: [13.661223034s] [13.661223034s] END Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.771117 4770 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.771083 4770 trace.go:236] Trace[415849776]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 11:31:18.517) (total time: 14253ms): Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[415849776]: ---"Objects listed" error: 14253ms (11:31:32.771) Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[415849776]: [14.253139493s] [14.253139493s] END Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.771175 4770 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 11:31:32 crc kubenswrapper[4770]: E1209 11:31:32.772772 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.772998 4770 trace.go:236] Trace[1170168062]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 11:31:19.443) (total time: 13329ms): Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[1170168062]: ---"Objects listed" error: 13329ms (11:31:32.772) Dec 09 11:31:32 crc kubenswrapper[4770]: Trace[1170168062]: [13.329652455s] [13.329652455s] END Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.773020 4770 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.787153 4770 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 09 11:31:32 crc kubenswrapper[4770]: I1209 11:31:32.795989 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.285217 4770 apiserver.go:52] "Watching apiserver" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.288072 4770 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.288320 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.288885 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.288894 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.289041 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.290045 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.290105 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.290184 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.290214 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.290284 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.291168 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.292522 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.292685 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.292744 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.293226 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.293418 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.295755 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.295769 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.296047 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.295998 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.321803 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.333042 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.348009 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.362093 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.374358 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.383217 4770 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.384498 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.400593 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.428341 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.431040 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9" exitCode=255 Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.431100 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9"} Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.438676 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.440518 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.440810 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.441386 4770 scope.go:117] "RemoveContainer" containerID="9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.442310 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.458398 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475228 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475284 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475310 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475332 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475356 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475406 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475430 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475452 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475480 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475526 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475549 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475571 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475595 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475654 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475679 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475701 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475702 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475726 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475730 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475806 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475940 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.475950 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476085 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476109 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476157 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476208 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476233 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476259 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476283 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476305 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476329 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476356 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476380 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476395 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476444 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476473 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476478 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476515 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476534 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476554 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476571 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476587 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476591 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476606 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476622 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476641 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476650 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476723 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479146 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479317 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479431 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479531 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479766 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479874 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479995 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480207 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480338 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480444 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480549 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480655 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480762 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480877 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481010 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481116 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481566 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481776 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481866 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482084 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482174 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482272 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482385 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482494 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482593 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482698 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482806 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.482936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483042 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483168 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483379 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483492 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483738 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483769 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483801 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483868 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483893 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476772 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.476970 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479031 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486449 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479005 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479155 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479322 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479566 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479781 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479836 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.479986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480169 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480181 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480421 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480560 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480603 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481175 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.480943 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.481570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483375 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483451 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483681 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483737 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.483991 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484273 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484338 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484438 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484454 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484561 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484588 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.484678 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485187 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485362 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.485383 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:31:33.98535799 +0000 UTC m=+19.226116509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485493 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485652 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485804 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486682 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.485843 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486005 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486100 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486118 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486165 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486345 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486920 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.486961 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487359 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487426 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487482 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487828 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487859 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.487986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.488271 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.488574 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.489027 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.489582 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.490248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.491150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.491439 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.491729 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.492007 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.492312 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.492485 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.492811 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.493121 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.493333 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.494755 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.495975 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.496017 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.496256 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.496678 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.497015 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.497167 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.497330 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.497452 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.497679 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.498035 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.498676 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.499017 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.499297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.499368 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.500239 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.501958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.502962 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.504421 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.505771 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509014 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509085 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509123 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509164 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509191 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509218 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509249 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509276 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509298 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509382 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509428 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509454 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509480 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509504 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509534 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.509562 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.510524 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511103 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511170 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511358 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511502 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511613 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.511983 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512194 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512565 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512744 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.512878 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.513151 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.518482 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.520143 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.524934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525050 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525084 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525112 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525135 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525161 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525212 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525234 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525264 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525293 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525317 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525349 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525401 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525458 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525488 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525513 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525540 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525568 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525591 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525617 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525644 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525671 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525695 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525723 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525749 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525772 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525801 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525828 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525851 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525923 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525953 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.525976 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526003 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526027 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526051 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526077 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526159 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526218 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526271 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526299 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526381 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526396 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526436 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.526764 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.527255 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.527968 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.528186 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.528497 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.527472 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.528515 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.529591 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.529743 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530052 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530212 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530227 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.528246 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530409 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530771 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.530967 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532211 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.531008 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532359 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532711 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532894 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532543 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532558 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.532996 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.529959 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533235 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533249 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533308 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533339 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533364 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533388 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533482 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533507 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533527 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533545 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533558 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533582 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533621 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533639 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533675 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533692 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533710 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533727 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533743 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533741 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533762 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533782 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533802 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533827 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533854 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533879 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533920 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533966 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533992 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534015 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534129 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534134 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534185 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534216 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534276 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534300 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534326 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534352 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534378 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534397 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534405 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534451 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534482 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534543 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534618 4770 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534635 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534649 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534663 4770 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534676 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534690 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534734 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534748 4770 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534761 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534773 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534786 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534800 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534812 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534826 4770 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534830 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534838 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534869 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534886 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534915 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534931 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.534958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.533812 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535235 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535544 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535632 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535649 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535726 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.535983 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536079 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536101 4770 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536119 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536155 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536174 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536244 4770 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536265 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536281 4770 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536295 4770 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536308 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536339 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536353 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536366 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536378 4770 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536392 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536404 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536414 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536425 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536435 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536447 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536456 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536465 4770 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536478 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536487 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536496 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536505 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536514 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536524 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536533 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536541 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536550 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536559 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536568 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536575 4770 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536584 4770 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536592 4770 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536601 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536610 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536620 4770 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536628 4770 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536819 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536832 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536840 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536850 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536858 4770 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536866 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536875 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536883 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536892 4770 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536931 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536942 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536951 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536960 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536969 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536978 4770 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536988 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.536999 4770 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537007 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537015 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537024 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537032 4770 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537042 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537049 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537058 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537068 4770 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537076 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537084 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537093 4770 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537355 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537360 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537486 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537514 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537809 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.537912 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538032 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538053 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538103 4770 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538145 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538164 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538182 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538198 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538216 4770 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538238 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538257 4770 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538274 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538287 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538301 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538301 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538317 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538333 4770 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538349 4770 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538362 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538375 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538388 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538401 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538416 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538432 4770 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538447 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538461 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538477 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538506 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538546 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538561 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538575 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538588 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538600 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538613 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538627 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538643 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538659 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538673 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538685 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538698 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538712 4770 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.538713 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.539154 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.539480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.539900 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.538726 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.540040 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:34.038801154 +0000 UTC m=+19.279559663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540321 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540384 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540320 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540462 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540479 4770 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540676 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540699 4770 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.540978 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541001 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541020 4770 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541058 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541072 4770 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541085 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541096 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541108 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541153 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541166 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541178 4770 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541190 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541324 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541324 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541338 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541681 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.541730 4770 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.542511 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.542874 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.543178 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.543483 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.543821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.544072 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.545059 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.545453 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.545544 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:34.04552451 +0000 UTC m=+19.286283029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.547069 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.547272 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.547348 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.550794 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.551353 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.552328 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.553236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.553943 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.556711 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.557236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.558730 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.560073 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.560281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.560623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562260 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562295 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562313 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562395 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:34.062367697 +0000 UTC m=+19.303126216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.562539 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562582 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562885 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562924 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:33 crc kubenswrapper[4770]: E1209 11:31:33.562991 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:34.062968712 +0000 UTC m=+19.303727421 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.563128 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.564369 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.567856 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.568098 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.568111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.571298 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.574340 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.578586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.579841 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.585330 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.592874 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.597104 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.602875 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.607029 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.611929 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.617983 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: W1209 11:31:33.623353 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-c7f1929ab478b281353a21d1f8c3b04ac9cdf1303bc4a5ac1e0cda37ef5a8d93 WatchSource:0}: Error finding container c7f1929ab478b281353a21d1f8c3b04ac9cdf1303bc4a5ac1e0cda37ef5a8d93: Status 404 returned error can't find the container with id c7f1929ab478b281353a21d1f8c3b04ac9cdf1303bc4a5ac1e0cda37ef5a8d93 Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.632268 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641608 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641695 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641707 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641717 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641725 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641734 4770 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641742 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641751 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641760 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641769 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641778 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641805 4770 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641796 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641815 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641885 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641896 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641936 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641945 4770 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641954 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641964 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641973 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641982 4770 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641991 4770 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642001 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642009 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.641998 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642018 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642095 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642110 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642125 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642140 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642152 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642166 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642179 4770 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642192 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642206 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642219 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642232 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642245 4770 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642256 4770 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642270 4770 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642283 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642296 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642308 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642320 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642333 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642347 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642360 4770 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642372 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642385 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642397 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.642411 4770 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.644778 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.656679 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.670808 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.904364 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 11:31:33 crc kubenswrapper[4770]: I1209 11:31:33.920400 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 11:31:33 crc kubenswrapper[4770]: W1209 11:31:33.932889 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e35833eadb90f0ac5f452f9d43f2e8cd69662895b006a015bd0d73bb562893c2 WatchSource:0}: Error finding container e35833eadb90f0ac5f452f9d43f2e8cd69662895b006a015bd0d73bb562893c2: Status 404 returned error can't find the container with id e35833eadb90f0ac5f452f9d43f2e8cd69662895b006a015bd0d73bb562893c2 Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.045549 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.045654 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.045688 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.045763 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.045820 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:35.045803552 +0000 UTC m=+20.286562071 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.045834 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:31:35.045827672 +0000 UTC m=+20.286586191 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.045929 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.046039 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:35.046018267 +0000 UTC m=+20.286776856 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.146825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.147124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147001 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147309 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147415 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147182 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147521 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147536 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147573 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:35.147559182 +0000 UTC m=+20.388317701 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:34 crc kubenswrapper[4770]: E1209 11:31:34.147675 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:35.147663615 +0000 UTC m=+20.388422134 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.436242 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.438447 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.439474 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.443786 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e35833eadb90f0ac5f452f9d43f2e8cd69662895b006a015bd0d73bb562893c2"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.445255 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.445441 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7db8e31d9580a0cc45806f70766e9bdbd195c1e31491657fcd72f4e29cc11d75"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.447789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.447926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.448022 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c7f1929ab478b281353a21d1f8c3b04ac9cdf1303bc4a5ac1e0cda37ef5a8d93"} Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.465552 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.487004 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.503015 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.518924 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.539158 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.554188 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.569108 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.581734 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.595489 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.607463 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.619829 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.634620 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.647643 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.661420 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.672989 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:34 crc kubenswrapper[4770]: I1209 11:31:34.685038 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:34Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.054521 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.054613 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.054643 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.054719 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.054791 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:31:37.054758633 +0000 UTC m=+22.295517162 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.054831 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:37.054820844 +0000 UTC m=+22.295579453 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.054962 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.055016 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:37.055006219 +0000 UTC m=+22.295764818 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.155375 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.155441 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155556 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155593 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155607 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155556 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155677 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155690 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155657 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:37.155641412 +0000 UTC m=+22.396399931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.155734 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:37.155724094 +0000 UTC m=+22.396482613 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.340028 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.340073 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.340181 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.340209 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.340314 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:35 crc kubenswrapper[4770]: E1209 11:31:35.340485 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.344186 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.345057 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.345934 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.346739 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.347522 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.348208 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.348957 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.349615 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.350419 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.351137 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.351766 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.352604 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.355107 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.355582 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.355952 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.356769 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.358214 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.359251 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.360466 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.361202 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.361773 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.362261 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.362935 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.363421 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.364193 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.364619 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.365203 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.365792 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.366354 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.367023 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.367557 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.369189 4770 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.369318 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.370647 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.370762 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.371401 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.371781 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.372854 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.373603 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.374226 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.374975 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.375632 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.376119 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.376693 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.378714 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.379395 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.379837 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.380466 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.381317 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.382297 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.382891 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.383474 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.384445 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.384862 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.385642 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.386943 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.387896 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.395962 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.406774 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.421313 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.434084 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.444207 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.972941 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.974200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.974254 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.974270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.974354 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.982063 4770 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.982493 4770 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.983624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.983658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.983669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.983691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:35 crc kubenswrapper[4770]: I1209 11:31:35.983706 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:35Z","lastTransitionTime":"2025-12-09T11:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.002076 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.006642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.006726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.006743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.006762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.006775 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.025244 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.028713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.028737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.028745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.028757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.028766 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.048731 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.052409 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.052472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.052484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.052508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.052520 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.067980 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.072620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.072680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.072693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.072715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.072728 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.088502 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: E1209 11:31:36.088640 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.091336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.091400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.091413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.091432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.091444 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.194019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.194088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.194107 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.194132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.194147 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.296551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.296605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.296615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.296630 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.296642 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.399934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.400058 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.400072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.400090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.400102 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.454979 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.468460 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.481713 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.494490 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.502926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.502972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.502981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.502998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.503008 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.506153 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.520341 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.534006 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.545162 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.557833 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.605725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.605763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.605774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.605791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.605805 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.709146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.709198 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.709210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.709230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.709243 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.812889 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.812996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.813011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.813036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.813050 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.912992 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.918325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.918359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.918367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.918384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.918393 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:36Z","lastTransitionTime":"2025-12-09T11:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.924466 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.927378 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.928781 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.940251 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.953067 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.964631 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.978258 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:36 crc kubenswrapper[4770]: I1209 11:31:36.989877 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:36Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.002118 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.014929 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.020776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.020821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.020834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.020849 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.020859 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.026894 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.046984 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.059557 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.070383 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.070503 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.070583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.070593 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.070604 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:31:41.070581392 +0000 UTC m=+26.311339911 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.070747 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.070762 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:41.070713606 +0000 UTC m=+26.311472125 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.070814 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:41.070798118 +0000 UTC m=+26.311556657 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.073083 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.085607 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.099953 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.112957 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.123510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.123547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.123558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.123576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.123588 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.125162 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.137007 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:37Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.171692 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.171760 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171882 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171948 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171949 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171960 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171972 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.171986 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.172023 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:41.172006785 +0000 UTC m=+26.412765304 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.172044 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:41.172037065 +0000 UTC m=+26.412795584 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.226213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.226264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.226277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.226294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.226306 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.329017 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.329067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.329078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.329095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.329106 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.339331 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.339414 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.339350 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.339963 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.340147 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:37 crc kubenswrapper[4770]: E1209 11:31:37.340275 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.431432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.431473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.431485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.431500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.431511 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.533875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.533925 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.533934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.533947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.533958 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.635693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.635732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.635744 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.635759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.635769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.738372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.738417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.738428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.738443 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.738454 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.840271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.840309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.840320 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.840336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.840348 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.936856 4770 csr.go:261] certificate signing request csr-krxj6 is approved, waiting to be issued Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.942748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.942789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.942797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.942817 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.942826 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:37Z","lastTransitionTime":"2025-12-09T11:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:37 crc kubenswrapper[4770]: I1209 11:31:37.955106 4770 csr.go:257] certificate signing request csr-krxj6 is issued Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.045424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.046038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.046119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.046191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.046248 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.097630 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vdpdw"] Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.098140 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.101024 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.101033 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.101519 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.101567 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.112183 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.128100 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.145359 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.149143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.149856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.149937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.149964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.149981 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.161989 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.176041 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.181573 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48b86990-8140-48c9-beaf-ef4bcbd26313-host\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.181653 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcqdx\" (UniqueName: \"kubernetes.io/projected/48b86990-8140-48c9-beaf-ef4bcbd26313-kube-api-access-kcqdx\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.181687 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/48b86990-8140-48c9-beaf-ef4bcbd26313-serviceca\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.190460 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.201276 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-s872h"] Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.201808 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.204832 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.204883 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.205782 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.218773 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.237629 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.250456 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.251612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.251651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.251663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.251681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.251693 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.263161 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282373 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgw6k\" (UniqueName: \"kubernetes.io/projected/313bc8c2-c09d-42ba-adbc-12698e1b26eb-kube-api-access-hgw6k\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282489 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48b86990-8140-48c9-beaf-ef4bcbd26313-host\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282525 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcqdx\" (UniqueName: \"kubernetes.io/projected/48b86990-8140-48c9-beaf-ef4bcbd26313-kube-api-access-kcqdx\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282534 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48b86990-8140-48c9-beaf-ef4bcbd26313-host\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282553 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/313bc8c2-c09d-42ba-adbc-12698e1b26eb-hosts-file\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.282583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/48b86990-8140-48c9-beaf-ef4bcbd26313-serviceca\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.284326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/48b86990-8140-48c9-beaf-ef4bcbd26313-serviceca\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.291259 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.311983 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcqdx\" (UniqueName: \"kubernetes.io/projected/48b86990-8140-48c9-beaf-ef4bcbd26313-kube-api-access-kcqdx\") pod \"node-ca-vdpdw\" (UID: \"48b86990-8140-48c9-beaf-ef4bcbd26313\") " pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.318771 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.331115 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.343047 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.354157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.354195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.354205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.354222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.354235 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.371744 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.383103 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgw6k\" (UniqueName: \"kubernetes.io/projected/313bc8c2-c09d-42ba-adbc-12698e1b26eb-kube-api-access-hgw6k\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.383184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/313bc8c2-c09d-42ba-adbc-12698e1b26eb-hosts-file\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.383261 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/313bc8c2-c09d-42ba-adbc-12698e1b26eb-hosts-file\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.395537 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.409681 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vdpdw" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.415408 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgw6k\" (UniqueName: \"kubernetes.io/projected/313bc8c2-c09d-42ba-adbc-12698e1b26eb-kube-api-access-hgw6k\") pod \"node-resolver-s872h\" (UID: \"313bc8c2-c09d-42ba-adbc-12698e1b26eb\") " pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.423392 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.446953 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.456377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.456597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.456701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.456806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.456882 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.461535 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vdpdw" event={"ID":"48b86990-8140-48c9-beaf-ef4bcbd26313","Type":"ContainerStarted","Data":"ffc4a6676f1b6fa5e8ffbb13d93bd6ab5c11754404cd445b2d7e5a3c5f0cfc11"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.462524 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.481928 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.493496 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.514645 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-s872h" Dec 09 11:31:38 crc kubenswrapper[4770]: W1209 11:31:38.539439 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod313bc8c2_c09d_42ba_adbc_12698e1b26eb.slice/crio-2d384c5feef7a1abaa0471c466a8e0dbd87845d1d277a69b5664e5459bf75734 WatchSource:0}: Error finding container 2d384c5feef7a1abaa0471c466a8e0dbd87845d1d277a69b5664e5459bf75734: Status 404 returned error can't find the container with id 2d384c5feef7a1abaa0471c466a8e0dbd87845d1d277a69b5664e5459bf75734 Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.559465 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.559504 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.559515 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.559532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.559544 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.663331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.663397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.663410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.663454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.663470 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.766767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.766811 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.766821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.766838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.766851 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.869670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.869726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.869745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.869767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.869778 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.956250 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-09 11:26:37 +0000 UTC, rotation deadline is 2026-11-02 11:02:41.969158046 +0000 UTC Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.956309 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7871h31m3.012855906s for next certificate rotation Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.972076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.972115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.972125 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.972141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:38 crc kubenswrapper[4770]: I1209 11:31:38.972150 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:38Z","lastTransitionTime":"2025-12-09T11:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.064017 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qxvgc"] Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.064505 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.064712 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-q2lxs"] Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.065003 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.066204 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.066315 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.066803 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.067126 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.067159 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.067464 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.069376 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.069656 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.069777 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.070072 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zt2b2"] Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.071540 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.071576 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.071892 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-shdpv"] Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.074038 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.074840 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.075102 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.075134 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.080228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.080269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.080280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.080298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.080311 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.083586 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.084640 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.087255 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.087494 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.087649 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.087806 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.095736 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.126814 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.176106 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.182099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.182139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.182151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.182167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.182181 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190470 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-conf-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190511 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqdtt\" (UniqueName: \"kubernetes.io/projected/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-kube-api-access-zqdtt\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190533 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cnibin\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190568 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190591 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-os-release\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18e4a331-a41b-419a-9a35-f9fb57ff38f1-proxy-tls\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190671 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190694 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190716 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-daemon-config\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190808 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190945 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190979 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.190997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-socket-dir-parent\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191033 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191048 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e4a331-a41b-419a-9a35-f9fb57ff38f1-mcd-auth-proxy-config\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191098 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191113 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191150 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-k8s-cni-cncf-io\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191186 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztphj\" (UniqueName: \"kubernetes.io/projected/18e4a331-a41b-419a-9a35-f9fb57ff38f1-kube-api-access-ztphj\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191227 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191281 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-etc-kubernetes\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191307 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cnibin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191323 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-system-cni-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191357 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/18e4a331-a41b-419a-9a35-f9fb57ff38f1-rootfs\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191376 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-system-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-bin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191413 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191429 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191446 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-multus-certs\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191463 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191480 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmjvf\" (UniqueName: \"kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191521 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-netns\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191550 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-multus\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191571 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191587 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191626 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191643 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cni-binary-copy\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191701 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-hostroot\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191753 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191796 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxhsw\" (UniqueName: \"kubernetes.io/projected/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-kube-api-access-kxhsw\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191812 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-os-release\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191828 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-kubelet\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.191876 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.195349 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.230221 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.246755 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.260316 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.280124 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.284034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.284077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.284088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.284105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.284116 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292629 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztphj\" (UniqueName: \"kubernetes.io/projected/18e4a331-a41b-419a-9a35-f9fb57ff38f1-kube-api-access-ztphj\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292657 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-k8s-cni-cncf-io\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292684 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-etc-kubernetes\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292711 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292736 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292762 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-system-cni-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292821 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cnibin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-etc-kubernetes\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292851 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292857 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-bin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-k8s-cni-cncf-io\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.292975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cnibin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293002 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293016 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-system-cni-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293022 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293064 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293057 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-bin\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293166 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293180 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/18e4a331-a41b-419a-9a35-f9fb57ff38f1-rootfs\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-system-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293264 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-multus-certs\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293286 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293312 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-netns\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293334 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-multus\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293356 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293378 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmjvf\" (UniqueName: \"kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293454 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293487 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-system-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293481 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293529 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-cni-multus\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293561 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-hostroot\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293576 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-multus-certs\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293585 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293602 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293619 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cni-binary-copy\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-run-netns\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293645 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-kubelet\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293671 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293691 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/18e4a331-a41b-419a-9a35-f9fb57ff38f1-rootfs\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293734 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293697 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293754 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-hostroot\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293764 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxhsw\" (UniqueName: \"kubernetes.io/projected/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-kube-api-access-kxhsw\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293821 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-os-release\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293863 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-conf-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293883 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqdtt\" (UniqueName: \"kubernetes.io/projected/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-kube-api-access-zqdtt\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293933 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cnibin\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293958 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293974 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-os-release\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.293791 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294004 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-os-release\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294025 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18e4a331-a41b-419a-9a35-f9fb57ff38f1-proxy-tls\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294028 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294046 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294064 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294094 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294115 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-daemon-config\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294165 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294188 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-socket-dir-parent\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294209 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-host-var-lib-kubelet\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294214 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294238 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294257 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e4a331-a41b-419a-9a35-f9fb57ff38f1-mcd-auth-proxy-config\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294609 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294717 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-cni-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294782 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cnibin\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294811 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-conf-dir\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294867 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e4a331-a41b-419a-9a35-f9fb57ff38f1-mcd-auth-proxy-config\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-socket-dir-parent\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.294240 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295084 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295122 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295144 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295234 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295276 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-os-release\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295521 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295796 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-multus-daemon-config\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.295793 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-cni-binary-copy\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.298032 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.298183 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18e4a331-a41b-419a-9a35-f9fb57ff38f1-proxy-tls\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.299530 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.310781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztphj\" (UniqueName: \"kubernetes.io/projected/18e4a331-a41b-419a-9a35-f9fb57ff38f1-kube-api-access-ztphj\") pod \"machine-config-daemon-qxvgc\" (UID: \"18e4a331-a41b-419a-9a35-f9fb57ff38f1\") " pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.318023 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqdtt\" (UniqueName: \"kubernetes.io/projected/4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86-kube-api-access-zqdtt\") pod \"multus-q2lxs\" (UID: \"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\") " pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.319439 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxhsw\" (UniqueName: \"kubernetes.io/projected/ab11cf58-3fcb-4d07-b532-d81db0beb2c3-kube-api-access-kxhsw\") pod \"multus-additional-cni-plugins-shdpv\" (UID: \"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\") " pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.320699 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.322168 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmjvf\" (UniqueName: \"kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf\") pod \"ovnkube-node-zt2b2\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.332987 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.339165 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.339209 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.339165 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:39 crc kubenswrapper[4770]: E1209 11:31:39.339278 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:39 crc kubenswrapper[4770]: E1209 11:31:39.339385 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:39 crc kubenswrapper[4770]: E1209 11:31:39.339545 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.344634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.357809 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.374412 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.385155 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.386245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.386276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.386286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.386303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.386320 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.387604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.397442 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-q2lxs" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.397798 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: W1209 11:31:39.398436 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18e4a331_a41b_419a_9a35_f9fb57ff38f1.slice/crio-2a28519cd16b9063754514f657ac8c6f263eaa9113ec7799a889d34d2baae1ff WatchSource:0}: Error finding container 2a28519cd16b9063754514f657ac8c6f263eaa9113ec7799a889d34d2baae1ff: Status 404 returned error can't find the container with id 2a28519cd16b9063754514f657ac8c6f263eaa9113ec7799a889d34d2baae1ff Dec 09 11:31:39 crc kubenswrapper[4770]: W1209 11:31:39.407998 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eb284c3_ad1f_44f7_9bf7_f1a477d4dc86.slice/crio-e538c4f844230a61e8fd7b8026ca1b4ca54f2468ebd192944c0556f294a7bfba WatchSource:0}: Error finding container e538c4f844230a61e8fd7b8026ca1b4ca54f2468ebd192944c0556f294a7bfba: Status 404 returned error can't find the container with id e538c4f844230a61e8fd7b8026ca1b4ca54f2468ebd192944c0556f294a7bfba Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.408669 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.413146 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.419996 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-shdpv" Dec 09 11:31:39 crc kubenswrapper[4770]: W1209 11:31:39.421006 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaa774d4_1a7d_4731_a47b_d1d97c88869e.slice/crio-8668c6db0a39095b91845c2a1d06c8f8cab407f6c46944fc5539a7c9721fce2e WatchSource:0}: Error finding container 8668c6db0a39095b91845c2a1d06c8f8cab407f6c46944fc5539a7c9721fce2e: Status 404 returned error can't find the container with id 8668c6db0a39095b91845c2a1d06c8f8cab407f6c46944fc5539a7c9721fce2e Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.430257 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: W1209 11:31:39.440081 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab11cf58_3fcb_4d07_b532_d81db0beb2c3.slice/crio-e5283357f8ccdbe0ab820788f73ce11bf76f41414ffedcc2522c1c97dbf64949 WatchSource:0}: Error finding container e5283357f8ccdbe0ab820788f73ce11bf76f41414ffedcc2522c1c97dbf64949: Status 404 returned error can't find the container with id e5283357f8ccdbe0ab820788f73ce11bf76f41414ffedcc2522c1c97dbf64949 Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.444550 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.466886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerStarted","Data":"e5283357f8ccdbe0ab820788f73ce11bf76f41414ffedcc2522c1c97dbf64949"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.468471 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"8668c6db0a39095b91845c2a1d06c8f8cab407f6c46944fc5539a7c9721fce2e"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.470180 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-s872h" event={"ID":"313bc8c2-c09d-42ba-adbc-12698e1b26eb","Type":"ContainerStarted","Data":"350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.470204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-s872h" event={"ID":"313bc8c2-c09d-42ba-adbc-12698e1b26eb","Type":"ContainerStarted","Data":"2d384c5feef7a1abaa0471c466a8e0dbd87845d1d277a69b5664e5459bf75734"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.470781 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.473988 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerStarted","Data":"e538c4f844230a61e8fd7b8026ca1b4ca54f2468ebd192944c0556f294a7bfba"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.475287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"2a28519cd16b9063754514f657ac8c6f263eaa9113ec7799a889d34d2baae1ff"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.476437 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vdpdw" event={"ID":"48b86990-8140-48c9-beaf-ef4bcbd26313","Type":"ContainerStarted","Data":"2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.485693 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.489449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.489513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.489530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.489547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.489559 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.498277 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.510173 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.519894 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.532783 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.544572 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.564324 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.577361 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.587763 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.591604 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.591637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.591648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.591664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.591675 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.607541 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.619738 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.634238 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.647814 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.661497 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.672560 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.683410 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.694387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.694446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.694461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.694480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.694493 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.696139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.708641 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.731658 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.743919 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.757176 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.772254 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.797033 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.797077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.797089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.797105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.797114 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.899705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.899748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.899760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.899776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:39 crc kubenswrapper[4770]: I1209 11:31:39.899786 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:39Z","lastTransitionTime":"2025-12-09T11:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.002450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.002771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.002780 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.002795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.002809 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.104756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.104798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.104809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.104827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.104838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.207394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.207453 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.207502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.207550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.207595 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.310336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.310364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.310373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.310387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.310398 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.412357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.412391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.412402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.412415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.412425 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.481220 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.481274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.483366 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0" exitCode=0 Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.483426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.484702 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" exitCode=0 Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.484742 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.487123 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerStarted","Data":"440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.506616 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.516525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.517051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.517267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.517399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.517488 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.522624 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.541454 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.554557 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.564579 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.578224 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.594697 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.605422 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.620503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.620994 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.621064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.621171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.621248 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.626819 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.647245 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.662560 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.675825 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.690979 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.706997 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.723757 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.724346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.724383 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.724408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.724428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.724442 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.742781 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.755656 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.789782 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.818255 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.830080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.830321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.830403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.830495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.830576 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.847876 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.862145 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.874888 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.886894 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.910637 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.932262 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.933638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.933679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.933691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.933709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.933721 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:40Z","lastTransitionTime":"2025-12-09T11:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.948490 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.966179 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.982029 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:40 crc kubenswrapper[4770]: I1209 11:31:40.993465 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.008171 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.036321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.036419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.036712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.036995 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.037226 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.115588 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.115716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.115776 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:31:49.115754798 +0000 UTC m=+34.356513317 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.115816 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.115817 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.115863 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:49.11585379 +0000 UTC m=+34.356612309 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.115924 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.115960 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:49.115949982 +0000 UTC m=+34.356708501 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.139450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.139500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.139511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.139528 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.139540 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.216670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.216740 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216849 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216865 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216861 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216929 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216943 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216994 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:49.216978495 +0000 UTC m=+34.457737014 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.216876 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.217048 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:31:49.217033196 +0000 UTC m=+34.457791715 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.242554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.242587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.242598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.242613 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.242623 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.340095 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.340150 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.340148 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.340287 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.340376 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:41 crc kubenswrapper[4770]: E1209 11:31:41.340570 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.344692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.344744 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.344787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.344809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.344824 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.447015 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.447065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.447077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.447095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.447106 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.492065 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682" exitCode=0 Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.492159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.497914 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.497971 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.497983 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.497993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.498003 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.498013 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.508303 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.520634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.548462 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.549517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.549551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.549561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.549575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.549587 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.563328 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.580053 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.593187 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.606072 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.619487 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.629668 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.639637 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.652297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.652331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.652341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.652356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.652365 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.654156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.677848 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.694732 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.710612 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.724636 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:41Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.754993 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.755038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.755047 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.755062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.755074 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.858369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.858429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.858441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.858458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.858476 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.961162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.961446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.961542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.961644 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:41 crc kubenswrapper[4770]: I1209 11:31:41.961735 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:41Z","lastTransitionTime":"2025-12-09T11:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.064643 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.064930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.064939 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.064951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.064960 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.167652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.167682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.167690 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.167703 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.167712 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.270494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.270559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.270569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.270583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.270594 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.372983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.373092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.373114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.373135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.373147 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.475841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.475885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.475893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.475922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.475932 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.503319 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.503152 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579" exitCode=0 Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.520415 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.533626 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.553121 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.566491 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.577983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.578022 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.578032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.578049 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.578060 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.581969 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.593282 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.606677 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.621636 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.636253 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.650752 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.663661 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.677000 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.681242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.681286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.681297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.681320 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.681332 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.687410 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.699550 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.718716 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:42Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.783991 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.784034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.784045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.784062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.784073 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.886311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.886355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.886363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.886378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.886387 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.989736 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.989780 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.989790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.989803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:42 crc kubenswrapper[4770]: I1209 11:31:42.989812 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:42Z","lastTransitionTime":"2025-12-09T11:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.091765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.091800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.091808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.091821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.091830 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.194506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.194548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.194561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.194579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.194592 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.297303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.297371 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.297381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.297397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.297407 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.339205 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.339372 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.339581 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:43 crc kubenswrapper[4770]: E1209 11:31:43.339576 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:43 crc kubenswrapper[4770]: E1209 11:31:43.339660 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:43 crc kubenswrapper[4770]: E1209 11:31:43.339733 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.399708 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.399759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.399769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.399788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.399799 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.502637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.502694 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.502705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.502722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.502735 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.508973 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0" exitCode=0 Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.509061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.513982 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.530552 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.550075 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.566815 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.580224 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.590643 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.603537 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.604730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.604762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.604775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.604795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.604809 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.616121 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.627266 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.645624 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.661594 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.674857 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.687857 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.702814 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.707026 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.707061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.707071 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.707086 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.707096 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.716442 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.730836 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:43Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.809670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.809724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.809739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.809757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.809769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.912612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.912650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.912660 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.912675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:43 crc kubenswrapper[4770]: I1209 11:31:43.912686 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:43Z","lastTransitionTime":"2025-12-09T11:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.015464 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.015567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.015585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.016082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.016345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.120747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.120808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.120821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.120843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.121206 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.223955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.224004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.224015 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.224032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.224043 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.327342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.327409 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.327424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.327448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.327465 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.431075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.431402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.431411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.431426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.431435 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.521078 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a" exitCode=0 Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.521121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.534007 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.534065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.534078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.534104 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.534117 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.547051 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.567204 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.578974 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.592765 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.604864 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.620559 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.633668 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.637237 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.637269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.637279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.637294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.637305 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.646508 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.658791 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.672401 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.685552 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.700500 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.720185 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.734699 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.739521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.739559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.739570 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.739588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.739600 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.749231 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:44Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.842519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.842558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.842569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.842587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.842595 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.944977 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.945016 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.945042 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.945058 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:44 crc kubenswrapper[4770]: I1209 11:31:44.945070 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:44Z","lastTransitionTime":"2025-12-09T11:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.047541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.047585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.047598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.047615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.047627 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.150633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.150682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.150693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.150710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.150721 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.191359 4770 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.253434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.253474 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.253482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.253497 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.253506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.340031 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.340069 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:45 crc kubenswrapper[4770]: E1209 11:31:45.340148 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:45 crc kubenswrapper[4770]: E1209 11:31:45.340228 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.340248 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:45 crc kubenswrapper[4770]: E1209 11:31:45.340337 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.356186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.356247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.356258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.356277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.356290 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.367099 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.381445 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.394111 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.406321 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.419140 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.437196 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.452013 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.458774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.458818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.458830 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.458848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.458859 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.465957 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.477872 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.489644 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.501708 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.527503 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.528072 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab11cf58-3fcb-4d07-b532-d81db0beb2c3" containerID="da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4" exitCode=0 Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.528113 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerDied","Data":"da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.566210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.566427 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.566519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.566587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.566645 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.579990 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.597209 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.611635 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.625135 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.635671 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.657413 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.669357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.669402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.669414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.669429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.669441 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.670861 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.685370 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.700346 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.713476 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.727593 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.742228 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.762985 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.772400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.772479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.772515 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.772532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.772542 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.777841 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.792387 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.813515 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.829018 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.838614 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.874820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.874871 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.874882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.874922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.874936 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.977721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.977772 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.977781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.977796 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:45 crc kubenswrapper[4770]: I1209 11:31:45.977805 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:45Z","lastTransitionTime":"2025-12-09T11:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.080811 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.080848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.080857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.080870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.080881 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.182814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.183128 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.183220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.183301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.183382 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.261157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.261199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.261210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.261227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.261238 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.273653 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.277736 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.277777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.277788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.277806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.277820 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.290626 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.294055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.294090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.294100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.294113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.294123 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.311321 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.315103 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.315132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.315141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.315171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.315181 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.328167 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.331529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.331569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.331578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.331597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.331606 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.344167 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: E1209 11:31:46.344281 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.345963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.346000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.346010 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.346025 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.346036 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.449027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.449099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.449110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.449127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.449138 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.535585 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" event={"ID":"ab11cf58-3fcb-4d07-b532-d81db0beb2c3","Type":"ContainerStarted","Data":"e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.539429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.539881 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.541440 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.553717 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.554661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.554712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.554723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.554741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.554753 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.563516 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.570626 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.576553 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.596186 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.608203 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.620939 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.632250 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.645749 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.658302 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.658347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.658363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.658378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.658389 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.661285 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.675091 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.693493 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.710008 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.726519 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.737336 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.750485 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.761420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.761472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.761485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.761506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.761524 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.770476 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.786104 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.801485 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.820474 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.832436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.846286 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.858298 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.863468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.863510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.863524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.863539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.863551 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.876157 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.892810 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.908260 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.920754 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.931152 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.943593 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.954615 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.966211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.966258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.966270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.966286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.966297 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:46Z","lastTransitionTime":"2025-12-09T11:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.972644 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:46 crc kubenswrapper[4770]: I1209 11:31:46.983023 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:46Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.006779 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.019452 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.036073 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.049505 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.061470 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.068944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.068980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.068990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.069005 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.069015 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.077262 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.091634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.105269 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.116208 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.134890 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.147666 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.160189 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.170770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.170822 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.170836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.170852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.170865 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.172987 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.184849 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.200021 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.221678 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:47Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.274096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.274144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.274154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.274172 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.274186 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.339476 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.339525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.339476 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:47 crc kubenswrapper[4770]: E1209 11:31:47.339645 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:47 crc kubenswrapper[4770]: E1209 11:31:47.339706 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:47 crc kubenswrapper[4770]: E1209 11:31:47.339781 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.376410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.376457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.376468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.376485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.376499 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.478532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.479977 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.479996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.480011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.480021 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.542578 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.582478 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.582519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.582530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.582549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.582560 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.685121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.685173 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.685185 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.685200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.685212 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.787348 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.787400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.787410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.787424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.787432 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.889535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.889579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.889592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.889617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.889629 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.992099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.992140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.992152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.992171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:47 crc kubenswrapper[4770]: I1209 11:31:47.992182 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:47Z","lastTransitionTime":"2025-12-09T11:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.095146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.095188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.095199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.095215 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.095226 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.197241 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.197274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.197288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.197311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.197351 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.299600 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.299639 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.299648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.299671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.299681 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.401552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.401597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.401609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.401626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.401637 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.508689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.508935 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.509005 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.509067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.509171 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.545288 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.611973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.612251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.612561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.612763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.612997 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.715329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.715380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.715391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.715412 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.715425 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.817798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.817840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.817851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.817867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.817878 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.920338 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.920375 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.920386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.920403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:48 crc kubenswrapper[4770]: I1209 11:31:48.920415 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:48Z","lastTransitionTime":"2025-12-09T11:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.022533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.022598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.022617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.022639 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.022652 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.125089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.125130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.125139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.125156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.125169 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.202944 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.203043 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.203075 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.203146 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.203164 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:32:05.203135904 +0000 UTC m=+50.443894423 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.203194 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.203213 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:05.203194306 +0000 UTC m=+50.443952895 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.203266 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:05.203253957 +0000 UTC m=+50.444012556 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.228122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.228178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.228189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.228206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.228402 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.304476 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.304543 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304671 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304690 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304702 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304707 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304755 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:05.304738631 +0000 UTC m=+50.545497150 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304756 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304781 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.304849 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:05.304827113 +0000 UTC m=+50.545585692 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.330548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.330587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.330599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.330616 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.330627 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.339147 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.339191 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.339191 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.339290 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.339377 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:49 crc kubenswrapper[4770]: E1209 11:31:49.339462 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.433533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.433587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.433595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.433611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.433622 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.536194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.536232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.536241 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.536255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.536264 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.639333 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.639375 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.639385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.639400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.639410 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.741518 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.741887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.741925 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.741950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.741971 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.844764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.844833 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.844845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.844862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.844873 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.947656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.947704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.947715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.947731 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:49 crc kubenswrapper[4770]: I1209 11:31:49.947743 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:49Z","lastTransitionTime":"2025-12-09T11:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.050355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.050407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.050423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.050452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.050463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.152606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.152651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.152662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.152680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.152692 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.255782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.255841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.255857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.255875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.255888 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.358715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.358768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.358783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.358799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.358811 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.462013 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.462056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.462075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.462093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.462104 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.553215 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/0.log" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.557309 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae" exitCode=1 Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.557412 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.558552 4770 scope.go:117] "RemoveContainer" containerID="b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.563971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.564019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.564034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.564051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.564067 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.575778 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.589385 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.601634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.613372 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.626606 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.643978 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.666130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.666162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.666173 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.666186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.666195 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.668055 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.682846 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.694546 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.706271 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.716148 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.729729 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.744146 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.756816 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.768444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.768485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.768496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.768510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.768520 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.775687 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.811188 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb"] Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.811562 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.813123 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.813304 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.823287 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.837694 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.851084 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.865842 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.872865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.872931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.872944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.872963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.872975 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.885798 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.902492 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.916884 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.919536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.919611 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c867d\" (UniqueName: \"kubernetes.io/projected/88ce1c3d-e028-4084-a153-33840b5e93e8-kube-api-access-c867d\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.919684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.919708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88ce1c3d-e028-4084-a153-33840b5e93e8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.930625 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.943421 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.953786 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.966651 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.975322 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.975364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.975376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.975394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.975407 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:50Z","lastTransitionTime":"2025-12-09T11:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:50 crc kubenswrapper[4770]: I1209 11:31:50.982721 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.000573 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:50Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.012748 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:51Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.020171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c867d\" (UniqueName: \"kubernetes.io/projected/88ce1c3d-e028-4084-a153-33840b5e93e8-kube-api-access-c867d\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.020243 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.020264 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88ce1c3d-e028-4084-a153-33840b5e93e8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.020293 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.020916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.021175 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88ce1c3d-e028-4084-a153-33840b5e93e8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.023426 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:51Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.025480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88ce1c3d-e028-4084-a153-33840b5e93e8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.037586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c867d\" (UniqueName: \"kubernetes.io/projected/88ce1c3d-e028-4084-a153-33840b5e93e8-kube-api-access-c867d\") pod \"ovnkube-control-plane-749d76644c-hjqqb\" (UID: \"88ce1c3d-e028-4084-a153-33840b5e93e8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.042530 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:51Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.078299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.078358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.078374 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.078394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.078406 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.124804 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.180706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.180756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.180765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.180779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.180788 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: W1209 11:31:51.233181 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88ce1c3d_e028_4084_a153_33840b5e93e8.slice/crio-6e153ff02b4f8e76111d182f1078f74195bca666184c61420b07561d86d683d9 WatchSource:0}: Error finding container 6e153ff02b4f8e76111d182f1078f74195bca666184c61420b07561d86d683d9: Status 404 returned error can't find the container with id 6e153ff02b4f8e76111d182f1078f74195bca666184c61420b07561d86d683d9 Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.283314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.283348 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.283359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.283374 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.283385 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.321782 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.339286 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.339330 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:51 crc kubenswrapper[4770]: E1209 11:31:51.339451 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.339560 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:51 crc kubenswrapper[4770]: E1209 11:31:51.339672 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:51 crc kubenswrapper[4770]: E1209 11:31:51.339841 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.386202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.386236 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.386245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.386259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.386269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.489058 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.489095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.489112 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.489127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.489138 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.564715 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" event={"ID":"88ce1c3d-e028-4084-a153-33840b5e93e8","Type":"ContainerStarted","Data":"6e153ff02b4f8e76111d182f1078f74195bca666184c61420b07561d86d683d9"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.594689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.594742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.594752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.594778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.594796 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.697916 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.697964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.697976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.698000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.698013 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.801582 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.801628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.801637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.801655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.801666 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.904332 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.904384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.904395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.904417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:51 crc kubenswrapper[4770]: I1209 11:31:51.904431 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:51Z","lastTransitionTime":"2025-12-09T11:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.006799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.006874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.006890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.006940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.006966 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.109773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.109826 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.109838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.109855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.109866 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.212001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.212039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.212050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.212066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.212079 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.314762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.314806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.314816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.314831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.314840 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.417231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.417268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.417278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.417295 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.417305 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.519438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.519490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.519501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.519520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.519532 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.569068 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/0.log" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.571618 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.572127 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.573354 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" event={"ID":"88ce1c3d-e028-4084-a153-33840b5e93e8","Type":"ContainerStarted","Data":"717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.588505 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.601667 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.616395 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8pmjn"] Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.616888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: E1209 11:31:52.616968 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.622414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.622468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.622483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.622506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.622521 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.669753 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.686663 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.704132 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.718061 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.725808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.725885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.725910 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.725928 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.725940 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.745135 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvtt\" (UniqueName: \"kubernetes.io/projected/95971ae2-d7fe-4259-b041-9b6373d8a51c-kube-api-access-4zvtt\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.745419 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.753094 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.778173 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.797963 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.824733 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.828744 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.829013 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.829100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.829183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.829269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.846576 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvtt\" (UniqueName: \"kubernetes.io/projected/95971ae2-d7fe-4259-b041-9b6373d8a51c-kube-api-access-4zvtt\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.847010 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.846853 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: E1209 11:31:52.847140 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:52 crc kubenswrapper[4770]: E1209 11:31:52.847488 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:31:53.347450641 +0000 UTC m=+38.588209160 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.863705 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.874610 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvtt\" (UniqueName: \"kubernetes.io/projected/95971ae2-d7fe-4259-b041-9b6373d8a51c-kube-api-access-4zvtt\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.893181 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.907958 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.924297 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.933441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.933477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.933487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.933505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.933517 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:52Z","lastTransitionTime":"2025-12-09T11:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.943953 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.957127 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.970729 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:52 crc kubenswrapper[4770]: I1209 11:31:52.990608 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:52Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.004065 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.017708 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.029945 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.035442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.035524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.035538 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.035557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.035568 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.040181 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.056329 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.068323 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.088004 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.099832 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.108788 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.119628 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.131161 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.138300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.138359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.138371 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.138385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.138395 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.146273 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.159312 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.177438 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.241237 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.241279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.241288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.241303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.241312 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.339206 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.339369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.339436 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.339366 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.339206 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.339525 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.343093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.343128 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.343139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.343154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.343166 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.351287 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.351476 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.351567 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:31:54.351543527 +0000 UTC m=+39.592302106 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.445634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.445685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.445693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.445706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.445716 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.549798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.549892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.549936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.549966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.549986 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.578064 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/1.log" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.578647 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/0.log" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.580814 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348" exitCode=1 Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.580878 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.580938 4770 scope.go:117] "RemoveContainer" containerID="b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.582707 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" event={"ID":"88ce1c3d-e028-4084-a153-33840b5e93e8","Type":"ContainerStarted","Data":"83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.582760 4770 scope.go:117] "RemoveContainer" containerID="6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348" Dec 09 11:31:53 crc kubenswrapper[4770]: E1209 11:31:53.583017 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.598658 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.621161 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.635440 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652241 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652271 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.652393 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.667154 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.678591 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.692919 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.709558 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.722635 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.750269 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.754919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.754963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.754974 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.754992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.755006 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.765067 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.777800 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.793149 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.806872 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.820325 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.837870 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.853242 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.857518 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.857566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.857591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.857610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.857623 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.872803 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.888929 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.903549 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.920527 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.934362 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.947661 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.960259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.960301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.960311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.960326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.960336 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:53Z","lastTransitionTime":"2025-12-09T11:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.963442 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.978321 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:53 crc kubenswrapper[4770]: I1209 11:31:53.991727 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:53Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.003063 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.015958 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.029291 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.052041 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.063144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.063206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.063230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.063248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.063261 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.067083 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.079784 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.100643 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b184fa5643822b2552e5f25724ed9b57037f6295f7cc4cdeef331b6be689f6ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:49Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:31:48.887413 6056 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 11:31:48.887468 6056 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 11:31:48.887483 6056 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 11:31:48.887527 6056 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 11:31:48.887537 6056 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 11:31:48.887555 6056 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:48.887560 6056 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:48.887573 6056 factory.go:656] Stopping watch factory\\\\nI1209 11:31:48.887596 6056 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:48.887608 6056 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 11:31:48.887614 6056 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 11:31:48.887621 6056 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 11:31:48.887627 6056 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 11:31:48.887633 6056 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 11:31:48.887641 6056 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.114759 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.165615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.165654 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.165664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.165680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.165691 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.268168 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.268212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.268225 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.268246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.268259 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.339443 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:54 crc kubenswrapper[4770]: E1209 11:31:54.339648 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.360658 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:54 crc kubenswrapper[4770]: E1209 11:31:54.360807 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:54 crc kubenswrapper[4770]: E1209 11:31:54.360892 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:31:56.360872086 +0000 UTC m=+41.601630605 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.370070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.370114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.370127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.370144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.370153 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.472921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.472964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.472973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.472991 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.473011 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.575290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.575349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.575359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.575380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.575390 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.588655 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/1.log" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.593921 4770 scope.go:117] "RemoveContainer" containerID="6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348" Dec 09 11:31:54 crc kubenswrapper[4770]: E1209 11:31:54.594121 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.611010 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.623112 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.632051 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.643683 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.653131 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.668971 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.680144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.680245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.680261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.680285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.680305 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.683384 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.695544 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.709410 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.723423 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.739687 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.753106 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.771600 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.783802 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.783859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.783873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.783944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.783963 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.790932 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.817998 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.835995 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.850609 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:54Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.887661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.887720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.887734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.887755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.887768 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.990389 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.990449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.990460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.990479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:54 crc kubenswrapper[4770]: I1209 11:31:54.990491 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:54Z","lastTransitionTime":"2025-12-09T11:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.093461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.093562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.093577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.093601 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.093617 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.196251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.196296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.196307 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.196325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.196337 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.298689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.298762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.298775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.298791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.298800 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.342050 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:55 crc kubenswrapper[4770]: E1209 11:31:55.342167 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.342445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:55 crc kubenswrapper[4770]: E1209 11:31:55.342499 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.342548 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:55 crc kubenswrapper[4770]: E1209 11:31:55.342594 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.357854 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.372839 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.383278 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.393474 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.401064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.401102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.401115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.401130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.401142 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.407856 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.420369 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.441674 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.454202 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.467030 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.480297 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.490272 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.503431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.503481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.503493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.503510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.503528 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.508709 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.522369 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.537200 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.547731 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.565072 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.577823 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.606614 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.606661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.606671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.606689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.606700 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.709339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.709371 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.709380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.709393 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.709402 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.812152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.812198 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.812206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.812220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.812230 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.914553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.914590 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.914602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.914618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:55 crc kubenswrapper[4770]: I1209 11:31:55.914630 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:55Z","lastTransitionTime":"2025-12-09T11:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.016964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.017009 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.017027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.017047 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.017059 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.119500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.119532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.119540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.119553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.119562 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.221763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.221825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.221839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.221858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.221879 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.325032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.325124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.325140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.325162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.325177 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.339113 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.339309 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.382789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.382997 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.383110 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:32:00.383079124 +0000 UTC m=+45.623837703 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.427821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.427859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.427871 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.427887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.427917 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.531315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.531357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.531366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.531383 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.531395 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.582220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.582267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.582276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.582297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.582309 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.596284 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:56Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.600171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.600231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.600242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.600263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.600275 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.614444 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:56Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.621407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.621469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.621479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.621495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.621506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.632503 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:56Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.636004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.636048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.636060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.636077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.636089 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.647331 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:56Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.651018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.651060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.651073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.651089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.651101 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.662123 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:31:56Z is after 2025-08-24T17:21:41Z" Dec 09 11:31:56 crc kubenswrapper[4770]: E1209 11:31:56.662242 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.663533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.663565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.663573 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.663596 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.663607 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.766245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.766287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.766297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.766309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.766319 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.868196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.868276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.868291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.868309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.868321 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.971221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.971263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.971273 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.971289 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:56 crc kubenswrapper[4770]: I1209 11:31:56.971299 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:56Z","lastTransitionTime":"2025-12-09T11:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.074321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.074373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.074386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.074404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.074417 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.177596 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.177672 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.177685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.177709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.177727 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.280240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.280278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.280291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.280311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.280323 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.339603 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.339658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.339620 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:57 crc kubenswrapper[4770]: E1209 11:31:57.339748 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:57 crc kubenswrapper[4770]: E1209 11:31:57.339814 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:57 crc kubenswrapper[4770]: E1209 11:31:57.339873 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.383568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.383609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.383619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.383632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.383641 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.486392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.486431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.486442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.486458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.486468 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.588727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.588777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.588791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.588810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.588822 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.691258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.691301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.691312 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.691333 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.691341 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.793792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.793832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.793840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.793854 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.793865 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.895838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.895883 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.895894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.895923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.895943 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.998159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.998197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.998213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.998228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:57 crc kubenswrapper[4770]: I1209 11:31:57.998238 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:57Z","lastTransitionTime":"2025-12-09T11:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.100316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.100352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.100370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.100387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.100398 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.202623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.202668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.202678 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.202692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.202703 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.306290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.306360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.306369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.306385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.306394 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.340034 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:31:58 crc kubenswrapper[4770]: E1209 11:31:58.340185 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.408328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.408369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.408380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.408402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.408413 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.510003 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.510046 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.510076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.510096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.510107 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.612773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.612832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.612843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.612862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.612874 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.715580 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.715619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.715631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.715651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.715664 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.818761 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.818829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.818840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.818858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.818868 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.920692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.920740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.920749 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.920763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:58 crc kubenswrapper[4770]: I1209 11:31:58.920772 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:58Z","lastTransitionTime":"2025-12-09T11:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.022860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.022917 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.022929 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.022944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.022957 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.125972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.126056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.126081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.126108 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.126136 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.228954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.229016 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.229032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.229055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.229074 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.332887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.332950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.332959 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.332976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.332988 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.339092 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.339163 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:31:59 crc kubenswrapper[4770]: E1209 11:31:59.339222 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.339266 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:31:59 crc kubenswrapper[4770]: E1209 11:31:59.339318 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:31:59 crc kubenswrapper[4770]: E1209 11:31:59.339405 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.436867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.437324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.437333 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.437347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.437356 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.539893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.540006 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.540022 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.540036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.540045 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.642157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.642205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.642217 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.642297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.642321 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.744563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.744624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.744640 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.744663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.744681 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.847992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.848036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.848057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.848074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.848087 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.952048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.952116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.952131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.952154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:31:59 crc kubenswrapper[4770]: I1209 11:31:59.952170 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:31:59Z","lastTransitionTime":"2025-12-09T11:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.054998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.055034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.055051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.055070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.055083 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.158530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.158578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.158592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.158608 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.158617 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.262040 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.262089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.262105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.262123 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.262138 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.339542 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:00 crc kubenswrapper[4770]: E1209 11:32:00.339768 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.366248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.366339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.366393 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.366434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.366463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.424283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:00 crc kubenswrapper[4770]: E1209 11:32:00.424463 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:00 crc kubenswrapper[4770]: E1209 11:32:00.424548 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:32:08.424522975 +0000 UTC m=+53.665281524 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.469091 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.469137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.469156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.469179 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.469196 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.573073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.573132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.573151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.573174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.573192 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.675673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.675710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.675718 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.675732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.675741 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.777664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.777708 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.777720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.777736 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.777748 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.880232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.880282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.880293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.880310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.880323 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.982013 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.982055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.982063 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.982076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:00 crc kubenswrapper[4770]: I1209 11:32:00.982086 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:00Z","lastTransitionTime":"2025-12-09T11:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.084783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.084847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.084870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.084893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.084998 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.187070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.187113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.187122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.187136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.187146 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.289329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.289390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.289402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.289420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.289431 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.339757 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:01 crc kubenswrapper[4770]: E1209 11:32:01.340395 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.339995 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:01 crc kubenswrapper[4770]: E1209 11:32:01.340588 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.339805 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:01 crc kubenswrapper[4770]: E1209 11:32:01.340802 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.392018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.392149 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.392165 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.392183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.392193 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.494805 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.494856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.494868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.494883 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.494894 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.597536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.597646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.597668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.597697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.597710 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.699867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.699940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.699954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.699972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.699983 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.802414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.802479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.802491 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.802508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.802518 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.905347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.905396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.905408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.905427 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:01 crc kubenswrapper[4770]: I1209 11:32:01.905440 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:01Z","lastTransitionTime":"2025-12-09T11:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.007963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.008012 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.008023 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.008040 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.008051 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.110300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.110346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.110362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.110384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.110397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.212745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.212789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.212796 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.212813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.212821 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.315494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.315535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.315545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.315560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.315574 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.339756 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:02 crc kubenswrapper[4770]: E1209 11:32:02.339949 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.418421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.418470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.418480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.418495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.418506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.521231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.521308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.521324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.521345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.521358 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.623723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.623761 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.623769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.623783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.623793 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.726782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.726813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.726823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.726835 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.726843 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.829181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.829228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.829240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.829253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.829262 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.931574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.931620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.931629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.931643 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:02 crc kubenswrapper[4770]: I1209 11:32:02.931652 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:02Z","lastTransitionTime":"2025-12-09T11:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.033786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.033832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.033847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.033865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.033877 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.136268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.136304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.136313 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.136329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.136338 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.239335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.239382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.239392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.239410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.239424 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.340072 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.340201 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.340230 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:03 crc kubenswrapper[4770]: E1209 11:32:03.340356 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:03 crc kubenswrapper[4770]: E1209 11:32:03.340519 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:03 crc kubenswrapper[4770]: E1209 11:32:03.340700 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.342263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.342310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.342325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.342346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.342365 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.446277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.446318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.446326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.446341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.446350 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.548652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.548691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.548701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.548717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.548728 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.651511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.651559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.651568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.651586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.651598 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.754503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.754559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.754572 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.754593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.754606 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.857381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.857440 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.857453 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.857472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.857485 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.931402 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.945659 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.950466 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:03Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.960693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.960759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.960770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.960789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.960799 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:03Z","lastTransitionTime":"2025-12-09T11:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.966344 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:03Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:03 crc kubenswrapper[4770]: I1209 11:32:03.989540 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:03Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.005280 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.017187 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.028408 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.040276 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.052131 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.063317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.063367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.063376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.063391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.063401 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.066768 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.084550 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.098820 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.112135 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.133927 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.151137 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165239 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165270 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.165757 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.180134 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.190024 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:04Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.268048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.268106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.268121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.268138 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.268149 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.340216 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:04 crc kubenswrapper[4770]: E1209 11:32:04.340471 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.371396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.371444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.371456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.371475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.371487 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.473529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.473568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.473577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.473592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.473601 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.576655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.576707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.576717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.576737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.576749 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.679700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.679755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.679767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.679794 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.679805 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.782493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.782542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.782552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.782571 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.782583 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.885353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.885404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.885417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.885438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.885450 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.988517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.988574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.988584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.988610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:04 crc kubenswrapper[4770]: I1209 11:32:04.988630 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:04Z","lastTransitionTime":"2025-12-09T11:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.092654 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.092715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.092726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.092841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.092869 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.196452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.196521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.196534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.196554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.196567 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.275522 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.275649 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.275732 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.275783 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:32:37.275737723 +0000 UTC m=+82.516496242 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.275855 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:37.275834206 +0000 UTC m=+82.516592715 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.275953 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.276108 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.276173 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:37.276155554 +0000 UTC m=+82.516914083 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.299863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.299940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.299950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.299966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.299976 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.339542 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.339674 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.339819 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.340202 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.340324 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.341205 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.341484 4770 scope.go:117] "RemoveContainer" containerID="6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.357956 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.374642 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.376818 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377031 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377052 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.377053 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377064 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377239 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:37.377219037 +0000 UTC m=+82.617977746 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377115 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377273 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377282 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:05 crc kubenswrapper[4770]: E1209 11:32:05.377346 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:32:37.37732563 +0000 UTC m=+82.618084149 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.391265 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.404274 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.408175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.408288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.408302 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.408328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.408341 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.421165 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.433462 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.454037 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.469302 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.482813 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.498664 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.511012 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.511054 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.511069 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.511085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.511097 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.512565 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.527079 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.540616 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.554587 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.566703 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.585382 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.598561 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.611784 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.613263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.613292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.613302 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.613321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.613333 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.625669 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/1.log" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.627550 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.628093 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.642489 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.657148 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.671642 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.687876 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.706984 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.718511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.718579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.718592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.718610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.718624 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.721021 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.738206 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.765098 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.779427 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.797390 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.818274 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.820866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.820893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.820913 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.820929 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.820938 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.833436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.848503 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.863317 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.875567 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.901212 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.914104 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.924579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.924633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.924646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.924664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.924676 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:05Z","lastTransitionTime":"2025-12-09T11:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:05 crc kubenswrapper[4770]: I1209 11:32:05.935046 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:05Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.341073 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.341177 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.341220 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.341260 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.341258 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.341313 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.341354 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.341424 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.342771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.342809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.342818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.342832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.342841 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.358200 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.365230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.365270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.365282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.365298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.365310 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.381411 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.387156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.387206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.387217 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.387233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.387243 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.401729 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.408252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.408304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.408315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.408334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.408345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.422317 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.431707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.431760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.431775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.431801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.431816 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.447048 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:07 crc kubenswrapper[4770]: E1209 11:32:07.447209 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.449009 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.449064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.449078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.449095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.449108 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.551490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.551517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.551525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.551538 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.551548 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.653658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.653698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.653711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.653727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.653739 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.757078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.757141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.757152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.757170 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.757182 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.858762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.858810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.858824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.858843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.858856 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.961192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.961224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.961233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.961247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:07 crc kubenswrapper[4770]: I1209 11:32:07.961255 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:07Z","lastTransitionTime":"2025-12-09T11:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.063885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.063934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.063947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.063962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.063972 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.166289 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.166319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.166327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.166339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.166348 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.268799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.268834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.268842 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.268855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.268865 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.353300 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/2.log" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.354202 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/1.log" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.357162 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b" exitCode=1 Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.357204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.357257 4770 scope.go:117] "RemoveContainer" containerID="6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.358046 4770 scope.go:117] "RemoveContainer" containerID="516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b" Dec 09 11:32:08 crc kubenswrapper[4770]: E1209 11:32:08.358201 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.371407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.371754 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.371772 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.371795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.371810 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.376241 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.387592 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.400288 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.410885 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.421467 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.430933 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.446983 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:08 crc kubenswrapper[4770]: E1209 11:32:08.447216 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:08 crc kubenswrapper[4770]: E1209 11:32:08.447299 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:32:24.447281019 +0000 UTC m=+69.688039538 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.447362 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.458076 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.466861 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.474240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.474293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.474305 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.474323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.474335 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.484024 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.496829 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.507406 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.518514 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.528515 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.538321 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.555161 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.568988 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.576664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.576715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.576732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.576756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.576770 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.585869 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.679687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.679757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.679768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.679788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.679799 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.782692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.782730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.782740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.782755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.782765 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.885352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.885388 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.885400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.885418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.885437 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.987545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.987583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.987592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.987606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:08 crc kubenswrapper[4770]: I1209 11:32:08.987616 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:08Z","lastTransitionTime":"2025-12-09T11:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.090485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.090551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.090565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.090584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.090595 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.193094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.193146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.193155 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.193174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.193184 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.295274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.295323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.295331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.295345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.295354 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.339924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.339968 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.339924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:09 crc kubenswrapper[4770]: E1209 11:32:09.340051 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.339924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:09 crc kubenswrapper[4770]: E1209 11:32:09.340115 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:09 crc kubenswrapper[4770]: E1209 11:32:09.340156 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:09 crc kubenswrapper[4770]: E1209 11:32:09.340195 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.360991 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/2.log" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.398022 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.398062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.398071 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.398089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.398098 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.501067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.501140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.501153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.501177 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.501190 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.604681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.604707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.604715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.604727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.604738 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.707245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.707759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.707874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.708001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.708101 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.809997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.810035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.810045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.810061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.810070 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.912525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.912554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.912562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.912575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:09 crc kubenswrapper[4770]: I1209 11:32:09.912583 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:09Z","lastTransitionTime":"2025-12-09T11:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.015105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.015369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.015450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.015541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.015629 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.118497 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.118555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.118568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.118590 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.118603 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.221159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.221192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.221202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.221214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.221223 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.323862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.324240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.324345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.324457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.324549 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.427133 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.427170 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.427181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.427197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.427210 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.529498 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.529546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.529557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.529573 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.529584 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.632782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.633093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.633189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.633280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.633345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.736550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.736801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.736905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.737028 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.737311 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.839293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.839352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.839366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.839384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.839396 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.941559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.941791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.941852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.941915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:10 crc kubenswrapper[4770]: I1209 11:32:10.942011 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:10Z","lastTransitionTime":"2025-12-09T11:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.044196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.044239 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.044249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.044264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.044273 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.147064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.147288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.147379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.147489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.147584 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.249885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.250165 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.250280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.250379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.250464 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.339151 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:11 crc kubenswrapper[4770]: E1209 11:32:11.339504 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.339250 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:11 crc kubenswrapper[4770]: E1209 11:32:11.339723 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.339281 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.339205 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:11 crc kubenswrapper[4770]: E1209 11:32:11.340065 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:11 crc kubenswrapper[4770]: E1209 11:32:11.340969 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.352626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.352690 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.352704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.352722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.352734 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.456462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.456508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.456519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.456540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.456553 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.559563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.559602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.559614 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.559629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.559638 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.662808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.662865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.662882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.662905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.662944 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.764878 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.764937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.764951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.764968 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.764980 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.867253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.867292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.867303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.867317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.867325 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.970094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.970154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.970171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.970195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:11 crc kubenswrapper[4770]: I1209 11:32:11.970211 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:11Z","lastTransitionTime":"2025-12-09T11:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.072458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.072509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.072519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.072534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.072545 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.175602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.175666 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.175677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.175699 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.175713 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.280037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.280096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.280106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.280121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.280131 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.383219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.383253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.383271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.383288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.383301 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.485172 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.485204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.485212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.485225 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.485233 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.587822 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.587940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.587951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.587970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.587982 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.690652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.690698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.690709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.690723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.690734 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.793189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.793248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.793263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.793283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.793295 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.895253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.895310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.895323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.895342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:12 crc kubenswrapper[4770]: I1209 11:32:12.895352 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:12Z","lastTransitionTime":"2025-12-09T11:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.001700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.001749 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.001760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.001777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.001788 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.104560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.104609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.104623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.104640 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.104651 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.207084 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.207127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.207168 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.207187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.207199 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.309229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.309272 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.309285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.309302 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.309314 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.339687 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.339801 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.339890 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:13 crc kubenswrapper[4770]: E1209 11:32:13.339824 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:13 crc kubenswrapper[4770]: E1209 11:32:13.340032 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.340109 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:13 crc kubenswrapper[4770]: E1209 11:32:13.340196 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:13 crc kubenswrapper[4770]: E1209 11:32:13.340255 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.411993 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.412088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.412100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.412119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.412130 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.514401 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.514438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.514455 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.514472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.514485 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.616673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.616749 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.616767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.616789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.616803 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.721012 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.721056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.721253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.721272 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.721281 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.823508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.823536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.823545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.823559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.823568 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.926330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.926373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.926388 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.926406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:13 crc kubenswrapper[4770]: I1209 11:32:13.926417 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:13Z","lastTransitionTime":"2025-12-09T11:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.029996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.030057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.030070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.030093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.030106 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.132298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.132344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.132358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.132380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.132392 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.235134 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.235186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.235201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.235219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.235234 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.338255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.338565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.338658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.338746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.338857 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.441803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.442095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.442210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.442293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.442366 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.545979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.546047 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.546058 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.546077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.546089 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.648814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.649213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.649350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.649525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.649673 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.751869 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.752251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.752343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.752426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.752517 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.920677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.920715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.920724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.920739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:14 crc kubenswrapper[4770]: I1209 11:32:14.920751 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:14Z","lastTransitionTime":"2025-12-09T11:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.024220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.024263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.024271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.024286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.024295 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.127092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.127137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.127166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.127181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.127191 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.229863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.229933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.229945 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.229960 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.229971 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.332014 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.332285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.332352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.332418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.332492 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.339180 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.339259 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:15 crc kubenswrapper[4770]: E1209 11:32:15.339312 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.339179 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.339187 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:15 crc kubenswrapper[4770]: E1209 11:32:15.339440 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:15 crc kubenswrapper[4770]: E1209 11:32:15.339524 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:15 crc kubenswrapper[4770]: E1209 11:32:15.339615 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.357349 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.370711 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.388010 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.403047 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.417413 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.428415 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.434287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.434353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.434365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.434384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.434397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.439642 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.452920 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.464293 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.482736 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.494211 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.511047 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.523528 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.535770 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.537500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.537623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.537660 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.537677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.537689 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.546582 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.566468 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6790ca09029ce2aea0b193e31c64ecea1cfeadfe1a9d9f9ca835d47b6865e348\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"1] Creating *factory.egressNode crc took: 20.951139ms\\\\nI1209 11:31:52.845692 6206 factory.go:1336] Added *v1.Node event handler 7\\\\nI1209 11:31:52.845743 6206 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1209 11:31:52.846186 6206 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 11:31:52.846255 6206 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 11:31:52.846293 6206 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-service-ca-operator/metrics\\\\\\\"}\\\\nI1209 11:31:52.846306 6206 services_controller.go:360] Finished syncing service metrics on namespace openshift-service-ca-operator for network=default : 2.547493ms\\\\nI1209 11:31:52.846323 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:31:52.846331 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 11:31:52.846413 6206 factory.go:656] Stopping watch factory\\\\nI1209 11:31:52.846427 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:31:52.846461 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:31:52.846472 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 11:31:52.846481 6206 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:31:52.846540 6206 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.577303 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.590198 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.639806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.639861 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.639874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.639892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.639927 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.742317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.742385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.742397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.742415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.742427 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.845083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.845439 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.845508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.845533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.845551 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.948424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.948471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.948483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.948501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:15 crc kubenswrapper[4770]: I1209 11:32:15.948512 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:15Z","lastTransitionTime":"2025-12-09T11:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.051216 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.051256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.051268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.051283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.051295 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.153800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.153856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.153868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.153887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.153919 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.256545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.256586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.256595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.256612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.256623 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.359380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.359760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.359784 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.359812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.359831 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.465148 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.465473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.465566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.465638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.465706 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.568595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.568639 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.568649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.568663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.568672 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.671232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.671267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.671278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.671297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.671313 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.773415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.773494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.773508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.773530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.773545 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.876116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.876163 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.876176 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.876194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.876206 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.978319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.978362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.978372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.978386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:16 crc kubenswrapper[4770]: I1209 11:32:16.978397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:16Z","lastTransitionTime":"2025-12-09T11:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.080820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.080868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.080878 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.080894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.080920 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.183234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.183285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.183294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.183310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.183326 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.286472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.286533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.286550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.286576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.286592 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.340270 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.340477 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.340659 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.340722 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.340747 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.340879 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.341038 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.341114 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.388342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.388531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.388557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.388581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.388597 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.490996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.491073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.491098 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.491124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.491142 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.593704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.593745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.593757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.593773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.593784 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.697499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.697819 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.697838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.697855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.697866 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.710215 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.710259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.710268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.710283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.710294 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.723710 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.727648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.727692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.727704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.727720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.727732 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.742257 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.746643 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.746686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.746698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.746716 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.746729 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.763877 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.767611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.767668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.767679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.767697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.767710 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.781184 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.784713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.785057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.785147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.785230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.785311 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.798291 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:17 crc kubenswrapper[4770]: E1209 11:32:17.798455 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.800124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.800159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.800173 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.800189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.800201 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.903077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.903116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.903127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.903143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:17 crc kubenswrapper[4770]: I1209 11:32:17.903154 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:17Z","lastTransitionTime":"2025-12-09T11:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.005995 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.006266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.006422 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.006578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.006656 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.109007 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.109049 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.109062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.109076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.109086 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.211231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.211281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.211292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.211309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.211320 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.314171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.314211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.314219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.314234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.314243 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.417109 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.417154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.417165 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.417184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.417197 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.519947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.519986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.519998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.520019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.520035 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.621823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.621870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.621884 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.621924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.621936 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.724740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.724790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.724831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.724851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.724867 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.827492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.827553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.827569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.827593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.827608 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.930145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.930441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.930517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.930586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:18 crc kubenswrapper[4770]: I1209 11:32:18.930646 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:18Z","lastTransitionTime":"2025-12-09T11:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.033403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.033449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.033462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.033479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.033493 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.135752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.136073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.136157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.136266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.136345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.238948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.238994 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.239007 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.239024 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.239036 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.339749 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.339770 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:19 crc kubenswrapper[4770]: E1209 11:32:19.339858 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:19 crc kubenswrapper[4770]: E1209 11:32:19.339935 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.340025 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.340070 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:19 crc kubenswrapper[4770]: E1209 11:32:19.340782 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:19 crc kubenswrapper[4770]: E1209 11:32:19.340974 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.346418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.346469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.346490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.346506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.346517 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.451833 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.451869 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.451879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.451893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.451920 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.555127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.555212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.555231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.555259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.555274 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.658144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.658191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.658202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.658220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.658232 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.760389 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.760430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.760442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.760461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.760474 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.864774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.864900 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.864932 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.864952 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.864967 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.967688 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.967733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.967742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.967761 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:19 crc kubenswrapper[4770]: I1209 11:32:19.967772 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:19Z","lastTransitionTime":"2025-12-09T11:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.071431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.071477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.071490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.071506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.071516 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.174855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.174941 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.174957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.174976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.174988 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.277461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.277503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.277512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.277527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.277541 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.379876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.379955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.379967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.379983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.379994 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.482274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.482331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.482342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.482360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.482373 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.584561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.584601 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.584612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.584627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.584639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.688012 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.688106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.688123 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.688156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.688178 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.790733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.790774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.790784 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.790801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.790812 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.893617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.893679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.893690 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.893710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.893724 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.996023 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.996072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.996085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.996103 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:20 crc kubenswrapper[4770]: I1209 11:32:20.996115 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:20Z","lastTransitionTime":"2025-12-09T11:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.099586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.099629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.099646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.099662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.099676 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.202222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.202262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.202275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.202292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.202304 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.305325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.305388 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.305400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.305417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.305428 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.339744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.339853 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.340040 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.340064 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:21 crc kubenswrapper[4770]: E1209 11:32:21.340197 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:21 crc kubenswrapper[4770]: E1209 11:32:21.340429 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:21 crc kubenswrapper[4770]: E1209 11:32:21.340592 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.340661 4770 scope.go:117] "RemoveContainer" containerID="516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b" Dec 09 11:32:21 crc kubenswrapper[4770]: E1209 11:32:21.340730 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:21 crc kubenswrapper[4770]: E1209 11:32:21.340888 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.355820 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.367326 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.389679 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408306 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408376 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.408572 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.425277 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.438436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.454511 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.467139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.478998 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.493405 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.504374 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.510916 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.510968 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.510976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.510993 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.511006 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.519309 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.539582 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.555382 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.571950 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.585613 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.594818 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.606532 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:21Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.613505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.613549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.613562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.613578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.613589 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.716773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.716827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.716841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.716859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.716872 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.820326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.820373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.820385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.820406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.820419 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.923135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.923174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.923186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.923203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:21 crc kubenswrapper[4770]: I1209 11:32:21.923215 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:21Z","lastTransitionTime":"2025-12-09T11:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.025007 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.025044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.025054 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.025067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.025076 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.127037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.127080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.127092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.127108 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.127122 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.229423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.229472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.229480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.229495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.229506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.331942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.331990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.332002 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.332026 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.332041 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.434625 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.434671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.434685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.434701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.434712 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.537435 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.537501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.537526 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.537546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.537561 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.641118 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.641170 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.641184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.641205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.641216 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.744224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.744266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.744282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.744298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.744309 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.846762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.846809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.846820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.846855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.846867 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.949379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.949428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.949443 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.949466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:22 crc kubenswrapper[4770]: I1209 11:32:22.949478 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:22Z","lastTransitionTime":"2025-12-09T11:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.052169 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.052223 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.052236 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.052257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.052269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.154459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.154503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.154513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.154530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.154541 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.256881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.256938 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.256947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.256961 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.256971 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.339306 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.339356 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.339315 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.339315 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:23 crc kubenswrapper[4770]: E1209 11:32:23.339481 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:23 crc kubenswrapper[4770]: E1209 11:32:23.339578 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:23 crc kubenswrapper[4770]: E1209 11:32:23.339663 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:23 crc kubenswrapper[4770]: E1209 11:32:23.339718 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.359038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.359093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.359103 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.359121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.359136 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.462370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.462404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.462412 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.462426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.462435 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.564646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.564696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.564707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.564723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.564734 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.667156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.667203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.667213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.667227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.667237 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.769882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.769935 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.769948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.769965 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.769976 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.871910 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.871969 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.871986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.872004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.872014 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.973917 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.973950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.973958 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.973971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:23 crc kubenswrapper[4770]: I1209 11:32:23.973979 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:23Z","lastTransitionTime":"2025-12-09T11:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.076451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.076487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.076495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.076511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.076522 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.179256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.179326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.179340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.179355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.179366 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.281323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.281367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.281378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.281394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.281406 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.383823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.383876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.383885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.383924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.383937 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.486310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.486402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.486421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.486677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.486692 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.521274 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:24 crc kubenswrapper[4770]: E1209 11:32:24.521506 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:24 crc kubenswrapper[4770]: E1209 11:32:24.521622 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:32:56.521601764 +0000 UTC m=+101.762360283 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.588563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.588610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.588625 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.588643 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.588655 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.690582 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.690628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.690638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.690652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.690662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.793548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.793602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.793615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.793636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.793650 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.896221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.896503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.896579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.896670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:24 crc kubenswrapper[4770]: I1209 11:32:24.896745 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:24.999888 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:24.999957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:24.999970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:24.999988 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.000000 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:24Z","lastTransitionTime":"2025-12-09T11:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.102653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.102697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.102710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.102731 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.102745 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.205285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.205321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.205329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.205342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.205359 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.307144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.307198 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.307211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.307229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.307239 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.339947 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.340013 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.340080 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:25 crc kubenswrapper[4770]: E1209 11:32:25.340192 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.340214 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:25 crc kubenswrapper[4770]: E1209 11:32:25.340332 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:25 crc kubenswrapper[4770]: E1209 11:32:25.340447 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:25 crc kubenswrapper[4770]: E1209 11:32:25.340541 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.352634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.365042 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.376278 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.386928 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.403032 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.411798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.411851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.411860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.411874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.411883 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.414622 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.432866 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.446340 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.460456 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.475869 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.489245 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.503237 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.513928 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.514189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.514280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.514289 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.514363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.514492 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.527359 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.537965 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.555254 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.567917 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.579925 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.616781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.616825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.616837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.616853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.616863 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.720508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.720585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.720605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.720630 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.720649 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.824287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.824337 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.824347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.824364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.824375 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.927027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.927598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.927692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.927796 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:25 crc kubenswrapper[4770]: I1209 11:32:25.927886 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:25Z","lastTransitionTime":"2025-12-09T11:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.030750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.030804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.030816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.030834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.030845 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.132967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.132998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.133006 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.133020 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.133029 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.235534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.235578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.235593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.235610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.235623 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.337990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.338081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.338095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.338110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.338121 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.416663 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/0.log" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.416745 4770 generic.go:334] "Generic (PLEG): container finished" podID="4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86" containerID="440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111" exitCode=1 Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.416792 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerDied","Data":"440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.417353 4770 scope.go:117] "RemoveContainer" containerID="440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.429229 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.440890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.440992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.441008 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.441037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.441051 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.443654 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.458929 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.475877 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.492139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.505868 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.522974 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.535591 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.552119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.552166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.552181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.552197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.552208 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.608194 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.626222 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.641810 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.654032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.654065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.654072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.654085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.654096 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.658002 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.669504 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.685160 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.699333 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.712890 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.722487 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.745240 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.755946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.755992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.756000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.756013 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.756022 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.857711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.857740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.857748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.857760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.857769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.960470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.960561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.960577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.960597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:26 crc kubenswrapper[4770]: I1209 11:32:26.960615 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:26Z","lastTransitionTime":"2025-12-09T11:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.063564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.063622 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.063633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.063657 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.063669 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.175574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.175648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.175661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.175682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.175997 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.278396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.278454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.278468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.278490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.278502 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.339700 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.339761 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.339775 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:27 crc kubenswrapper[4770]: E1209 11:32:27.339960 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:27 crc kubenswrapper[4770]: E1209 11:32:27.340113 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:27 crc kubenswrapper[4770]: E1209 11:32:27.340344 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.340377 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:27 crc kubenswrapper[4770]: E1209 11:32:27.340560 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.382525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.382592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.382605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.382626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.382638 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.422607 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/0.log" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.422661 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerStarted","Data":"b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.437264 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.447848 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.460001 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.470436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.484601 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.485794 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.485840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.485847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.485918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.485938 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.497709 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.518867 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.531530 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.544182 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.562310 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.576495 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.588939 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.588985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.588999 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.589018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.589031 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.589337 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.602722 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.616084 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.629135 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.646313 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.657373 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.672324 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.691675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.691730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.691742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.691762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.691774 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.794999 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.795074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.795095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.795114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.795143 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.897705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.897800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.897816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.897843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:27 crc kubenswrapper[4770]: I1209 11:32:27.897863 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:27Z","lastTransitionTime":"2025-12-09T11:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.000072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.000114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.000124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.000138 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.000147 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.101751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.101796 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.101806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.101822 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.101837 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.116171 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:28Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.120793 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.120828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.120841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.120857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.120868 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.140218 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:28Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.145923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.145965 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.145976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.145994 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.146005 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.158714 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:28Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.162670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.162722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.162733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.162753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.162766 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.174336 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:28Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.178175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.178212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.178223 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.178237 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.178246 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.191693 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:28Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:28 crc kubenswrapper[4770]: E1209 11:32:28.191866 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.194050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.194092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.194103 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.194122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.194132 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.296759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.296798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.296807 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.296820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.296829 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.398662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.398701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.398710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.398725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.398737 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.501579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.501636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.501706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.501733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.501749 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.604244 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.604287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.604297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.604312 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.604323 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.706967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.707018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.707027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.707044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.707055 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.809986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.810040 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.810050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.810069 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.810084 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.913280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.913359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.913372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.913394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:28 crc kubenswrapper[4770]: I1209 11:32:28.913407 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:28Z","lastTransitionTime":"2025-12-09T11:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.016474 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.016535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.016555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.016577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.016591 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.119034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.119105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.119115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.119139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.119153 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.221716 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.221765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.221778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.221794 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.221826 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.323523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.323598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.323614 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.323634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.323645 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.339376 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.339416 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.339532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:29 crc kubenswrapper[4770]: E1209 11:32:29.339566 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.339425 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:29 crc kubenswrapper[4770]: E1209 11:32:29.339702 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:29 crc kubenswrapper[4770]: E1209 11:32:29.339827 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:29 crc kubenswrapper[4770]: E1209 11:32:29.339941 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.426672 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.426728 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.426738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.426755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.426767 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.529405 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.529453 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.529466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.529481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.529492 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.632719 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.632781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.632792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.632809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.632820 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.735769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.735801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.735812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.735828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.735839 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.838850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.838924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.838950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.838966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.838975 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.941401 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.941443 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.941451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.941467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:29 crc kubenswrapper[4770]: I1209 11:32:29.941479 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:29Z","lastTransitionTime":"2025-12-09T11:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.043862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.043924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.043943 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.043962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.043971 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.149350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.149494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.149521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.149550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.149581 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.254454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.254504 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.254516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.254614 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.254628 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.356942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.357025 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.357038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.357052 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.357063 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.460516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.460592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.460619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.460648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.460661 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.563366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.563441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.563453 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.563477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.563495 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.666513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.666571 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.666583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.666599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.666610 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.769578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.769617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.769625 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.769676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.769697 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.872198 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.872239 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.872251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.872267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.872279 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.974890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.974959 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.974967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.974980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:30 crc kubenswrapper[4770]: I1209 11:32:30.974991 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:30Z","lastTransitionTime":"2025-12-09T11:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.077560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.077612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.077624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.077643 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.077655 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.180722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.180783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.180792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.180811 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.180823 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.283848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.283921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.283936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.283958 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.283974 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.339484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.339534 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.339532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.339484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:31 crc kubenswrapper[4770]: E1209 11:32:31.339642 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:31 crc kubenswrapper[4770]: E1209 11:32:31.339746 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:31 crc kubenswrapper[4770]: E1209 11:32:31.339822 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:31 crc kubenswrapper[4770]: E1209 11:32:31.339861 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.386390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.386464 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.386480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.386503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.386518 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.489150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.489190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.489201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.489217 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.489228 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.591576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.591618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.591627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.591641 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.591651 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.694439 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.694480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.694490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.694510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.694521 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.797155 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.797192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.797200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.797214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.797225 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.899838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.899886 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.899915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.899933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:31 crc kubenswrapper[4770]: I1209 11:32:31.899945 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:31Z","lastTransitionTime":"2025-12-09T11:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.002549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.002598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.002609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.002628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.002639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.104953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.104992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.105003 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.105019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.105030 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.208239 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.208513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.208588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.208669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.208744 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.312036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.312080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.312092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.312113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.312126 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.414858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.415200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.415271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.415346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.415407 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.518257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.518341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.518350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.518370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.518383 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.625275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.625318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.625328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.625342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.625352 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.727486 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.727706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.727769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.727866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.727956 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.831028 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.831074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.831084 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.831101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.831113 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.934328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.934376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.934387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.934411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:32 crc kubenswrapper[4770]: I1209 11:32:32.934426 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:32Z","lastTransitionTime":"2025-12-09T11:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.036748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.036788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.036799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.036815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.036827 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.138935 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.138990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.139004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.139021 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.139032 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.241634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.241705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.241719 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.241742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.241759 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.339493 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.339584 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:33 crc kubenswrapper[4770]: E1209 11:32:33.339636 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.339709 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.339584 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:33 crc kubenswrapper[4770]: E1209 11:32:33.339817 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:33 crc kubenswrapper[4770]: E1209 11:32:33.339866 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:33 crc kubenswrapper[4770]: E1209 11:32:33.339928 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.343870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.343923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.343934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.343944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.343953 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.447183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.447234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.447247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.447263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.447274 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.550710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.550771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.550785 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.550808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.550823 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.654373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.654420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.654434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.654451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.654461 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.757174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.757203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.757211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.757228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.757237 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.860660 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.860734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.860746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.860766 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.861177 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.964606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.964663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.964674 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.964697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:33 crc kubenswrapper[4770]: I1209 11:32:33.964713 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:33Z","lastTransitionTime":"2025-12-09T11:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.067923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.067997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.068018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.068038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.068051 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.170093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.170131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.170149 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.170167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.170180 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.272867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.272936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.272955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.272971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.272983 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.375284 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.375328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.375341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.375364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.375375 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.478637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.478689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.478700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.478724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.478735 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.581462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.581501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.581512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.581527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.581537 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.683806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.683894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.683926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.683943 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.683954 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.786345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.786386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.786397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.786411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.786420 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.888757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.888806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.888817 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.888832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.888843 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.991366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.991412 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.991421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.991436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:34 crc kubenswrapper[4770]: I1209 11:32:34.991446 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:34Z","lastTransitionTime":"2025-12-09T11:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.094460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.094522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.094539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.094556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.094566 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.197921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.197969 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.197978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.197992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.198004 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.300411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.300463 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.300477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.300495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.300513 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.339604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.339656 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.339699 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.339816 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:35 crc kubenswrapper[4770]: E1209 11:32:35.339993 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:35 crc kubenswrapper[4770]: E1209 11:32:35.340131 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:35 crc kubenswrapper[4770]: E1209 11:32:35.340291 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:35 crc kubenswrapper[4770]: E1209 11:32:35.340401 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.357319 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.368444 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.389358 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.402557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.402612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.402629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.402650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.402665 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.409385 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.421714 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.435018 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.447768 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.460450 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.471712 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.488942 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.503091 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.505485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.505568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.505588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.505618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.505688 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.524682 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.541272 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.554253 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.567866 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.580617 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.594590 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.607506 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:35Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.608764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.608812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.608824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.608843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.608855 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.710868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.710922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.710931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.710946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.710956 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.813806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.813842 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.813858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.813875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.813887 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.916106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.916143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.916154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.916201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:35 crc kubenswrapper[4770]: I1209 11:32:35.916214 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:35Z","lastTransitionTime":"2025-12-09T11:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.019089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.019134 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.019144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.019161 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.019172 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.121697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.121748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.121760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.121778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.121793 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.223797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.223864 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.223879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.223918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.223931 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.327255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.327305 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.327317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.327335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.327347 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.339695 4770 scope.go:117] "RemoveContainer" containerID="516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.430427 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.430467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.430477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.430529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.430544 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.532361 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.532411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.532421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.532435 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.532443 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.636412 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.636455 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.636464 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.636478 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.636488 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.738618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.738675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.738687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.738704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.738717 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.840996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.841027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.841035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.841049 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.841058 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.943711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.943739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.943747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.943759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:36 crc kubenswrapper[4770]: I1209 11:32:36.943767 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:36Z","lastTransitionTime":"2025-12-09T11:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.046588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.046630 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.046646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.046662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.046673 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.149219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.149264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.149275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.149317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.149330 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.251307 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.251352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.251362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.251379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.251390 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.340206 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.340249 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.340365 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.340369 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.340400 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.340500 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.340579 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.340642 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.347717 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.347939 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.347978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.348081 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.348139 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:33:41.348122589 +0000 UTC m=+146.588881108 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.348202 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:41.348194141 +0000 UTC m=+146.588952660 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.348273 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.348305 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 11:33:41.348296774 +0000 UTC m=+146.589055293 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.353396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.353422 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.353430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.353444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.353453 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.448662 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.448739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.448929 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449144 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449158 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449212 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 11:33:41.449194913 +0000 UTC m=+146.689953432 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449473 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449535 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449551 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:37 crc kubenswrapper[4770]: E1209 11:32:37.449650 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 11:33:41.449625694 +0000 UTC m=+146.690384213 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.456479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.456526 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.456537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.456552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.456563 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.559419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.559458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.559470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.559484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.559495 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.661341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.661363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.661372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.661394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.661403 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.763997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.764037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.764046 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.764059 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.764069 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.865877 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.866011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.866036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.866063 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.866085 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.971268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.971301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.971308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.971321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:37 crc kubenswrapper[4770]: I1209 11:32:37.971329 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:37Z","lastTransitionTime":"2025-12-09T11:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.074556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.074594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.074604 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.074626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.074637 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.177463 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.177504 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.177514 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.177529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.177540 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.375498 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.375548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.375558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.375575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.375585 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.388008 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.463293 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/2.log" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.465671 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.466266 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.478086 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.478133 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.478145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.478162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.478174 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.479517 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e319a92-0d4f-42ed-ad9b-7b5c4661a785\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.493718 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.504232 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.507534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.507561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.507569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.507584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.507596 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.520976 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.524174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.524219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.524232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.524250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.524263 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.529608 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.536368 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.539714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.539764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.539776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.539791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.539805 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.540218 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.552955 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.555075 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.556507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.556546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.556558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.556574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.556585 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.568279 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.568403 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.574403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.574442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.574452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.574469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.574482 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.585465 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: E1209 11:32:38.585628 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.586658 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.587294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.587347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.587358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.587377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.587388 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.598791 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.609359 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.624185 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.636911 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.658481 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.674625 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.687597 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.689300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.689329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.689340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.689354 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.689372 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.700153 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.711651 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.723236 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.733100 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.794221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.794266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.794277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.794303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.794313 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.897880 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.897943 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.897956 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.897972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:38 crc kubenswrapper[4770]: I1209 11:32:38.897984 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:38Z","lastTransitionTime":"2025-12-09T11:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.000650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.000684 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.000692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.000707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.000716 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.102813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.102852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.102860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.102873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.102883 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.205223 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.205299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.205318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.205348 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.205365 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.307791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.307846 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.307859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.307875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.307887 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.340210 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:39 crc kubenswrapper[4770]: E1209 11:32:39.340436 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.340623 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.340699 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:39 crc kubenswrapper[4770]: E1209 11:32:39.340795 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.340709 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:39 crc kubenswrapper[4770]: E1209 11:32:39.340853 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:39 crc kubenswrapper[4770]: E1209 11:32:39.340984 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.410587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.410636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.410646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.410671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.410685 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.470267 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.471055 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/2.log" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.473539 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" exitCode=1 Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.473615 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.473765 4770 scope.go:117] "RemoveContainer" containerID="516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.474363 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:32:39 crc kubenswrapper[4770]: E1209 11:32:39.474589 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.486662 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.499839 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.510139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.513579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.513633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.513650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.513673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.513690 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.523155 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.536305 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.544554 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.560205 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.573156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.585138 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.597871 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.610156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.615752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.615787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.615795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.615808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.615817 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.621716 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.631635 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.640402 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e319a92-0d4f-42ed-ad9b-7b5c4661a785\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.651313 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.660699 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.676394 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://516d105a6aacf6a054438834e1bf19b75e3df04f24f44edaa7ef2dcb6303789b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:08Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:07Z is after 2025-08-24T17:21:41Z]\\\\nI1209 11:32:07.712660 6407 services_controller.go:444] Built service openshift-kube-controller-manager-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712669 6407 services_controller.go:445] Built service openshift-kube-controller-manager-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1209 11:32:07.712665 6407 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1209 11:32:07.712682 6407 services_controller.go:451] Built service openshift-kube-controller-manager-operator/metrics cluster-wide L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.743988 6792 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.744071 6792 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744164 6792 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744794 6792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:32:38.744841 6792 factory.go:656] Stopping watch factory\\\\nI1209 11:32:38.744859 6792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:32:38.759951 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1209 11:32:38.760011 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1209 11:32:38.760281 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:32:38.760326 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:32:38.760405 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.686507 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.696593 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:39Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.718228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.718255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.718263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.718276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.718286 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.821023 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.821423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.821567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.821709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.821829 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.924236 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.924481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.924553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.924624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:39 crc kubenswrapper[4770]: I1209 11:32:39.924700 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:39Z","lastTransitionTime":"2025-12-09T11:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.027580 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.027612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.027620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.027632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.027641 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.130071 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.130331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.130413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.130497 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.130565 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.233225 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.233844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.234000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.234101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.234181 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.337268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.337317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.337327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.337346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.337362 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.440491 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.440557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.440601 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.440626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.440641 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.479016 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.483341 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:32:40 crc kubenswrapper[4770]: E1209 11:32:40.483552 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.497630 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.510229 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.524347 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.537846 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.544205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.544240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.544249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.544268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.544284 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.551246 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.564837 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.581547 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.592421 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.611345 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.624378 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.639629 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.647996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.648433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.648570 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.648690 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.648786 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.653742 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.665719 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.683369 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.697796 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.707951 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e319a92-0d4f-42ed-ad9b-7b5c4661a785\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.719810 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.729766 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.747122 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.743988 6792 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.744071 6792 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744164 6792 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744794 6792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:32:38.744841 6792 factory.go:656] Stopping watch factory\\\\nI1209 11:32:38.744859 6792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:32:38.759951 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1209 11:32:38.760011 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1209 11:32:38.760281 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:32:38.760326 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:32:38.760405 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:40Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.751774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.751825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.751836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.751857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.751870 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.853508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.853566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.853578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.853591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.853617 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.956497 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.956554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.956567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.956585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:40 crc kubenswrapper[4770]: I1209 11:32:40.956596 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:40Z","lastTransitionTime":"2025-12-09T11:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.060067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.060107 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.060116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.060131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.060141 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.162800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.162840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.162848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.162863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.162874 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.265264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.265352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.265377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.265410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.265435 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.339696 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.339770 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.339744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.339729 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:41 crc kubenswrapper[4770]: E1209 11:32:41.339931 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:41 crc kubenswrapper[4770]: E1209 11:32:41.340039 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:41 crc kubenswrapper[4770]: E1209 11:32:41.340106 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:41 crc kubenswrapper[4770]: E1209 11:32:41.340192 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.367486 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.367514 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.367524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.367542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.367552 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.470036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.470323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.470458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.470557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.470639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.573683 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.573726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.573743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.573773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.573800 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.676399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.676434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.676444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.676459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.676470 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.778794 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.778837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.778850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.778866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.778880 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.883440 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.883487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.883496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.883511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.883525 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.986623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.986691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.986706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.986724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:41 crc kubenswrapper[4770]: I1209 11:32:41.986758 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:41Z","lastTransitionTime":"2025-12-09T11:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.089211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.089609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.089798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.090083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.090292 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.194963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.195332 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.195447 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.195547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.195636 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.298696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.299421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.299556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.299660 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.299749 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.402099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.402131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.402142 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.402159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.402170 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.504339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.504389 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.504399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.504414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.504424 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.606963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.606996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.607004 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.607019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.607028 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.709143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.709414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.709531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.709735 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.709889 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.813082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.813115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.813127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.813141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.813151 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.915649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.915973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.916068 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.916164 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:42 crc kubenswrapper[4770]: I1209 11:32:42.916239 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:42Z","lastTransitionTime":"2025-12-09T11:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.018727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.019002 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.019074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.019202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.019262 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.121392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.121421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.121429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.121442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.121452 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.223187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.223227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.223237 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.223252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.223262 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.325866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.325920 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.325930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.325945 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.325956 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.339488 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.339563 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.339649 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:43 crc kubenswrapper[4770]: E1209 11:32:43.339644 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.339807 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:43 crc kubenswrapper[4770]: E1209 11:32:43.339801 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:43 crc kubenswrapper[4770]: E1209 11:32:43.339875 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:43 crc kubenswrapper[4770]: E1209 11:32:43.339974 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.428825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.428868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.428880 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.428920 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.428932 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.531370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.531407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.531419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.531434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.531444 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.633671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.633919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.633993 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.634060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.634134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.736805 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.736853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.736871 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.736892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.736922 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.839606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.839914 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.840005 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.840114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.840185 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.942921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.942967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.942978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.942992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:43 crc kubenswrapper[4770]: I1209 11:32:43.943001 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:43Z","lastTransitionTime":"2025-12-09T11:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.045766 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.045816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.045827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.045841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.045883 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.148962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.149005 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.149015 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.149031 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.149041 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.251281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.251324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.251336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.251355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.251368 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.354266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.354322 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.354330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.354350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.354360 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.456501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.456537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.456546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.456563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.456572 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.558228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.558255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.558263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.558275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.558284 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.660804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.661146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.661259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.661359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.661442 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.764483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.764524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.764541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.764557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.764568 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.867338 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.867382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.867407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.867423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.867435 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.970340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.970379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.970391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.970407 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:44 crc kubenswrapper[4770]: I1209 11:32:44.970426 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:44Z","lastTransitionTime":"2025-12-09T11:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.072484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.072726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.072823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.072982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.073092 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.176136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.176181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.176189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.176203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.176215 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.279062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.279098 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.279109 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.279125 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.279136 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.339686 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.339730 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.339694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:45 crc kubenswrapper[4770]: E1209 11:32:45.339816 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.339858 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:45 crc kubenswrapper[4770]: E1209 11:32:45.340180 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:45 crc kubenswrapper[4770]: E1209 11:32:45.340184 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:45 crc kubenswrapper[4770]: E1209 11:32:45.340282 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.359615 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.743988 6792 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.744071 6792 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744164 6792 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744794 6792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:32:38.744841 6792 factory.go:656] Stopping watch factory\\\\nI1209 11:32:38.744859 6792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:32:38.759951 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1209 11:32:38.760011 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1209 11:32:38.760281 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:32:38.760326 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:32:38.760405 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.374785 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e319a92-0d4f-42ed-ad9b-7b5c4661a785\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.382066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.382110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.382121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.382137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.382148 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.391039 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.403088 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.415834 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.430064 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.443711 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.457111 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.466692 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.477065 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.485120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.485150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.485158 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.485171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.485180 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.487280 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.499671 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.511257 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.520166 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.531437 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.541588 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.561763 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.576432 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.588634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.588668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.588677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.588692 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.588704 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.589361 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:45Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.691926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.691966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.691976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.691990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.692001 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.794559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.794618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.794632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.794657 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.794674 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.897776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.897824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.897837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.897856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:45 crc kubenswrapper[4770]: I1209 11:32:45.897871 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:45Z","lastTransitionTime":"2025-12-09T11:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.000436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.000478 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.000488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.000502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.000515 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.102668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.102713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.102726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.102744 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.102756 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.205948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.206006 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.206021 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.206039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.206050 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.308458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.308520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.308537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.308558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.308573 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.411011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.411062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.411073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.411089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.411100 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.513781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.513824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.513840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.513861 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.513878 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.616095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.616139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.616149 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.616166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.616177 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.718775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.718816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.718826 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.718839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.718851 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.821783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.821856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.821882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.821950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.821981 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.923851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.923923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.923934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.923948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:46 crc kubenswrapper[4770]: I1209 11:32:46.923959 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:46Z","lastTransitionTime":"2025-12-09T11:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.026583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.026647 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.026662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.026682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.026696 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.128808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.128848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.128862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.128879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.128896 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.232066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.232194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.232224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.232251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.232277 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.335035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.335094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.335108 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.335124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.335134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.340138 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.340258 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:47 crc kubenswrapper[4770]: E1209 11:32:47.340253 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.340291 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.340142 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:47 crc kubenswrapper[4770]: E1209 11:32:47.340399 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:47 crc kubenswrapper[4770]: E1209 11:32:47.340463 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:47 crc kubenswrapper[4770]: E1209 11:32:47.340601 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.437882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.437942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.437952 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.437968 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.437978 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.540274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.540311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.540321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.540335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.540344 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.643477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.643521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.643533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.643550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.643561 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.746413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.746455 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.746471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.746490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.746503 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.849444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.849483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.849507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.849523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.849535 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.951698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.951732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.951740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.951753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:47 crc kubenswrapper[4770]: I1209 11:32:47.951763 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:47Z","lastTransitionTime":"2025-12-09T11:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.054077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.054135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.054153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.054169 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.054181 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.156659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.156698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.156705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.156720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.156729 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.261097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.261147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.261184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.261202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.261219 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.364056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.364100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.364111 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.364128 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.364140 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.466583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.467003 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.467017 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.467036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.467049 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.569688 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.569732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.569741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.569755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.569764 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.671953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.671991 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.672001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.672014 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.672023 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.774584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.775043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.775070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.775102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.775126 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.877989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.878048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.878060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.878079 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.878091 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.914026 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.914088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.914106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.914121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.914131 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.927884 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.931986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.932025 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.932043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.932061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.932072 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.946143 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.950778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.950832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.950848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.950863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.950874 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.964210 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.968961 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.969011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.969022 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.969041 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.969054 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.982084 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.985864 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.985949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.985964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.985981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:48 crc kubenswrapper[4770]: I1209 11:32:48.985992 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:48Z","lastTransitionTime":"2025-12-09T11:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.999838 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f08a8155-c275-4917-b646-2fe0ff409bb7\\\",\\\"systemUUID\\\":\\\"c341d7e1-5989-4099-a8b0-1ad7487a3271\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:48 crc kubenswrapper[4770]: E1209 11:32:48.999967 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.001084 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.001113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.001122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.001136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.001145 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.103540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.103586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.103598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.103613 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.103626 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.205711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.205750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.205764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.205778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.205788 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.308675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.308705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.308713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.308727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.308736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.340171 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:49 crc kubenswrapper[4770]: E1209 11:32:49.340328 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.340534 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:49 crc kubenswrapper[4770]: E1209 11:32:49.340600 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.340765 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:49 crc kubenswrapper[4770]: E1209 11:32:49.340846 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.341023 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:49 crc kubenswrapper[4770]: E1209 11:32:49.341094 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.411563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.411619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.411631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.411649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.411661 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.513418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.513451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.513462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.513477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.513486 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.615814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.615847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.615857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.615873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.615884 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.718788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.718832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.718843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.718860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.718874 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.822122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.822158 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.822171 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.822187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.822199 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.924955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.925005 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.925017 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.925034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:49 crc kubenswrapper[4770]: I1209 11:32:49.925045 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:49Z","lastTransitionTime":"2025-12-09T11:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.029019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.029280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.029398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.029493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.029577 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.136019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.136072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.136082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.136098 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.136107 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.238366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.238418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.238430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.238448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.238458 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.340256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.340383 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.340398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.340415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.340427 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.443815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.443877 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.443890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.443944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.443960 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.546847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.546923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.546937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.546953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.546964 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.649436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.649483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.649494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.649509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.649518 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.752372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.752438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.752456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.752482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:50 crc kubenswrapper[4770]: I1209 11:32:50.752497 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.856130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.856195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.856208 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.856248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.856264 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.960072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.960103 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.960112 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.960128 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:50.960140 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:50Z","lastTransitionTime":"2025-12-09T11:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.062859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.062952 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.062971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.062985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.062995 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.165377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.165458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.165473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.165494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.165506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.268009 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.268063 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.268074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.268088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.268097 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.339529 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.339562 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.339638 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:51 crc kubenswrapper[4770]: E1209 11:32:51.339763 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.339951 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:51 crc kubenswrapper[4770]: E1209 11:32:51.339969 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:51 crc kubenswrapper[4770]: E1209 11:32:51.340045 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:51 crc kubenswrapper[4770]: E1209 11:32:51.340116 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.371758 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.371810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.371821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.371841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.371854 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.475591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.475637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.475647 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.475668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.475681 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.578019 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.578082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.578094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.578114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.578134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.680581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.680620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.680628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.680642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.680653 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.782549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.782590 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.782632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.782650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.782662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.885206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.885250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.885262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.885278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.885289 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.988166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.988229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.988241 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.988255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:51 crc kubenswrapper[4770]: I1209 11:32:51.988265 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:51Z","lastTransitionTime":"2025-12-09T11:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.090594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.090629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.090637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.090652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.090661 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.192588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.192618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.192626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.192638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.192648 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.295378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.295432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.295442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.295456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.295467 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.397842 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.397915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.397927 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.397942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.397952 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.501268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.501349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.501361 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.501380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.501393 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.604284 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.604334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.604343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.604360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.604371 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.707442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.707505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.707517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.707548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.707560 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.810732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.810844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.810859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.810881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.810937 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.915008 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.915064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.915076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.915097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:52 crc kubenswrapper[4770]: I1209 11:32:52.915110 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:52Z","lastTransitionTime":"2025-12-09T11:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.018040 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.018105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.018118 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.018145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.018165 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.121140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.121190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.121201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.121219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.121232 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.225479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.225541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.225551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.225565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.225578 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.329175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.329223 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.329233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.329249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.329259 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.339718 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:53 crc kubenswrapper[4770]: E1209 11:32:53.339865 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.340105 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:53 crc kubenswrapper[4770]: E1209 11:32:53.340230 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.340283 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.340587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:53 crc kubenswrapper[4770]: E1209 11:32:53.340678 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:53 crc kubenswrapper[4770]: E1209 11:32:53.340838 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.340991 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:32:53 crc kubenswrapper[4770]: E1209 11:32:53.341150 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.431937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.431980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.432009 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.432025 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.432035 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.534149 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.534218 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.534229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.534266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.534281 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.637289 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.637324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.637334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.637350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.637362 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.739970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.740089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.740102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.740127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.740141 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.843279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.843332 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.843342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.843358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.843372 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.948166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.948235 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.948250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.948276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:53 crc kubenswrapper[4770]: I1209 11:32:53.948295 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:53Z","lastTransitionTime":"2025-12-09T11:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.051156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.051203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.051212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.051232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.051245 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.154672 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.154726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.154736 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.154757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.154768 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.257543 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.257606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.257619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.257642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.257656 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.360377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.360411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.360420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.360434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.360443 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.463043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.463102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.463111 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.463125 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.463134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.565255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.565296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.565304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.565317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.565328 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.667260 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.667319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.667336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.667356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.667372 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.769182 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.769221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.769231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.769245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.769253 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.873276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.873332 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.873343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.873366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.873379 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.975415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.975470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.975482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.975503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:54 crc kubenswrapper[4770]: I1209 11:32:54.975527 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:54Z","lastTransitionTime":"2025-12-09T11:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.078131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.078170 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.078178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.078192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.078202 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.180402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.180458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.180469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.180488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.180499 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.283870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.284231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.284307 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.284381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.284452 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.339254 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.339552 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.339595 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.339628 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:55 crc kubenswrapper[4770]: E1209 11:32:55.339700 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:55 crc kubenswrapper[4770]: E1209 11:32:55.340072 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:55 crc kubenswrapper[4770]: E1209 11:32:55.340289 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:55 crc kubenswrapper[4770]: E1209 11:32:55.340884 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.355614 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40d03a12-38b2-4eca-b925-3081e48c40d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ac34cde328cd7196bfd1aa4da7fa51b046c5265ae8bfece5de18fa7dbd3a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f4b3cdbcfdfe89cf09e886aa5dfa165af163cb7c06286fde8660e83de404ff4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c6bb274092f770fe6be44fc8ca3c8b4a8fdea4ae64855fd1915ddc4bd3d034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.368967 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d3da488a0850929249e55a330dc42f52052672e88b219e600e0de5a54897e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585c4c2d04fa606a992736c37c1cc7edfc5d1ba9f7ea3e369cdb09e4913f0359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.380475 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s872h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313bc8c2-c09d-42ba-adbc-12698e1b26eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://350aa37819cf3bb35a617761450c6b48f46f873dd1bb4156883662aeafec468b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgw6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s872h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.387727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.387790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.387803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.387821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.387830 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.396379 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-q2lxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:26Z\\\",\\\"message\\\":\\\"2025-12-09T11:31:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0\\\\n2025-12-09T11:31:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_64b8ea63-3d0e-4271-9384-f0e1f3e6bdd0 to /host/opt/cni/bin/\\\\n2025-12-09T11:31:41Z [verbose] multus-daemon started\\\\n2025-12-09T11:31:41Z [verbose] Readiness Indicator file check\\\\n2025-12-09T11:32:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:32:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqdtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-q2lxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.407922 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88ce1c3d-e028-4084-a153-33840b5e93e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://717b878ef038b4e8b3eea8394c28981d686a22e691062170dffad8248e8deb55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ca875039e1c710f7e20d895a19e72f1a25e341a2ba64dfe3e17247744ec765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c867d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hjqqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.427222 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebf98812-d103-4cae-8e68-917214795336\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95209c56378b862621d54a2a02fa7640c12b8fa2e4c5ce1b3a292e155ae820e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcaf59709a2d2eddcb608c13b715b16354e3b61f821269e3516a7eb4107b6fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd9a23152fa615807878b6f5576945f9dc8d39f3180337fa4363dd92cdb61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a849ded443230dcf882a71292994081cfd962738fa240e1ed94a5f6e9f44e6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fd8aad66047c6bae371b6ddcaf455b83f407191be990d8d3c46349f92d62b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e6401d08e038c7a43c52bf20653f13e50753405fc717f170049dd98d8bdb58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65f103784bba529915f640a653112478cf60de7e735cb41803c53597e041483b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a6f241929cb180ef517fae1e25d5828d745d9a9e5110b3f7e46dcdd7843073d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.442259 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ff5f962-1825-4fc0-881d-dd3f982e99a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW1209 11:31:32.772455 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 11:31:32.772608 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 11:31:32.774506 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1049977914/tls.crt::/tmp/serving-cert-1049977914/tls.key\\\\\\\"\\\\nI1209 11:31:33.171488 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 11:31:33.181516 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 11:31:33.181543 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 11:31:33.181572 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 11:31:33.181578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 11:31:33.186531 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 11:31:33.186627 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186662 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 11:31:33.186695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 11:31:33.186724 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 11:31:33.186748 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 11:31:33.186772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 11:31:33.186556 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 11:31:33.188604 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.451415 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vdpdw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48b86990-8140-48c9-beaf-ef4bcbd26313\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d4529b77ec5dc1a569a0920440a1a9f6df8f7278ccf8c496e3227f3cda26e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kcqdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vdpdw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.468495 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa774d4-1a7d-4731-a47b-d1d97c88869e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T11:32:38Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.743988 6792 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 11:32:38.744071 6792 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744164 6792 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 11:32:38.744794 6792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 11:32:38.744841 6792 factory.go:656] Stopping watch factory\\\\nI1209 11:32:38.744859 6792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 11:32:38.759951 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1209 11:32:38.760011 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1209 11:32:38.760281 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1209 11:32:38.760326 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 11:32:38.760405 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T11:32:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmjvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zt2b2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.479434 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e319a92-0d4f-42ed-ad9b-7b5c4661a785\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f90373229231686ff0dca2f8edc9d2b3d39c106eb87461d2f816dff804f9671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab0764831be19e83b35b5dced28452cff2456dc17264855b3793147b6bd20de6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.489991 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.490024 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.490036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.490052 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.490063 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.495553 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee515dd88ba31855f0fa6d02361744eb53a7b33a8aaface5a27ac55902e667c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.508964 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6c63af-4e6e-476e-b3d7-51102741853b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418f2a555bdc5c2385534a2e712364f800a0ac324c9e00f004e64ad1858eb76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3c9b618f25bc9f7b28c68691f23c59779b1faa39b5a6ec307a2eb12668a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3571b6277ac957cb7513003baca5f8571752e72af776bab046260b7745ace6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a818cd53d44d909dde4fb503dce4754292db2ccbebe6536a327c46264949aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.522771 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.538452 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://694fb2a19305c873109048f1f697032aabb1212e61cb9221639842ff50c71331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.553377 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e4a331-a41b-419a-9a35-f9fb57ff38f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b1ffe6d2fd4493f4f6e2142737f87ed7b05050c996f2e955f2f8d7e66a53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ztphj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qxvgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.570587 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-shdpv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab11cf58-3fcb-4d07-b532-d81db0beb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e9f7cf9d5ea2e96f9aac820bbbd78f3adc485639033d62645ba737a1793ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T11:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbe543ef509b030ed9cfeb39775543d66a767642d5b8d272444debd4358f5e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352f17359cdeb2d222654939ad1471c17edc628edbdcd3abce52fc69c9719682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e82350d8d8bf26750f54030abe7b9671b223ebc80933729b9afe1062baf5579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315c2907de368686b43bc30fb65cef3cee20c553a413de82bdd80afc0a4c70f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e1645f93a83b7a157009626f6befa46c32a2b5cb22ccf3b4db59be87d0e7f4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da33f36824b8a45d0e4bb5310f0ccfbba0b623abb23485e04d65df13bdd5a1b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T11:31:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T11:31:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxhsw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-shdpv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.583113 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95971ae2-d7fe-4259-b041-9b6373d8a51c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zvtt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T11:31:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8pmjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.592987 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.593067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.593076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.593092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.593147 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.598488 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.614594 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T11:31:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T11:32:55Z is after 2025-08-24T17:21:41Z" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.696916 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.696997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.697011 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.697032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.697044 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.799298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.799334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.799342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.799356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.799366 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.902028 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.902319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.902395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.902476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:55 crc kubenswrapper[4770]: I1209 11:32:55.902555 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:55Z","lastTransitionTime":"2025-12-09T11:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.005870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.005948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.005962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.005979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.005992 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.108736 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.109192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.109298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.109387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.109473 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.212194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.212230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.212240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.212253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.212263 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.315247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.315709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.315780 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.315847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.315934 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.419104 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.419622 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.419798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.419960 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.420061 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.523178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.523430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.523452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.523480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.523500 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.566591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:56 crc kubenswrapper[4770]: E1209 11:32:56.566861 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:56 crc kubenswrapper[4770]: E1209 11:32:56.567009 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs podName:95971ae2-d7fe-4259-b041-9b6373d8a51c nodeName:}" failed. No retries permitted until 2025-12-09 11:34:00.566991063 +0000 UTC m=+165.807749582 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs") pod "network-metrics-daemon-8pmjn" (UID: "95971ae2-d7fe-4259-b041-9b6373d8a51c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.626778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.626832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.626844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.626861 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.626872 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.729627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.729675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.729684 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.729701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.729711 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.832203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.832500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.832592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.832688 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.832769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.934840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.934873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.934885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.934915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:56 crc kubenswrapper[4770]: I1209 11:32:56.934932 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:56Z","lastTransitionTime":"2025-12-09T11:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.037382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.037454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.037462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.037479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.037489 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.139981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.140032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.140045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.140061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.140073 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.242888 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.242946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.242955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.242969 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.242978 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.340116 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:57 crc kubenswrapper[4770]: E1209 11:32:57.340245 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.340316 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.340334 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.340546 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:57 crc kubenswrapper[4770]: E1209 11:32:57.340607 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:57 crc kubenswrapper[4770]: E1209 11:32:57.340765 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:57 crc kubenswrapper[4770]: E1209 11:32:57.340931 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.344982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.345027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.345040 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.345056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.345068 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.447817 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.447860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.447870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.447886 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.447918 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.550682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.550729 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.550743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.550758 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.550769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.653471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.653797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.653870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.653966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.654036 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.756444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.756494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.756506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.756522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.756532 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.858414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.858682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.859027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.859130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.859197 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.962063 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.962331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.962408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.962520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:57 crc kubenswrapper[4770]: I1209 11:32:57.962618 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:57Z","lastTransitionTime":"2025-12-09T11:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.066131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.066476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.066577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.066689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.066976 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.169994 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.170035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.170046 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.170062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.170073 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.272856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.273116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.273226 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.273311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.273382 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.376214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.376268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.376287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.376310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.376327 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.480038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.480114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.480136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.480177 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.480196 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.583150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.583202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.583213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.583235 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.583248 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.685781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.685831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.685845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.685887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.685926 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.788196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.788229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.788237 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.788250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.788258 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.891058 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.891118 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.891132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.891147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.891158 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.993084 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.993152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.993161 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.993174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:58 crc kubenswrapper[4770]: I1209 11:32:58.993182 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:58Z","lastTransitionTime":"2025-12-09T11:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.008651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.008694 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.008710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.008725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.008736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T11:32:59Z","lastTransitionTime":"2025-12-09T11:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.054589 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b"] Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.055024 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.057506 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.057808 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.058056 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.059224 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.111886 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=86.111866775 podStartE2EDuration="1m26.111866775s" podCreationTimestamp="2025-12-09 11:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.111127616 +0000 UTC m=+104.351886135" watchObservedRunningTime="2025-12-09 11:32:59.111866775 +0000 UTC m=+104.352625294" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.112483 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=83.112474641 podStartE2EDuration="1m23.112474641s" podCreationTimestamp="2025-12-09 11:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.097364013 +0000 UTC m=+104.338122532" watchObservedRunningTime="2025-12-09 11:32:59.112474641 +0000 UTC m=+104.353233150" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.137995 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.137953635 podStartE2EDuration="1m26.137953635s" podCreationTimestamp="2025-12-09 11:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.125824793 +0000 UTC m=+104.366583312" watchObservedRunningTime="2025-12-09 11:32:59.137953635 +0000 UTC m=+104.378712154" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.151601 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-s872h" podStartSLOduration=81.151580844 podStartE2EDuration="1m21.151580844s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.149721787 +0000 UTC m=+104.390480296" watchObservedRunningTime="2025-12-09 11:32:59.151580844 +0000 UTC m=+104.392339363" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.163344 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-q2lxs" podStartSLOduration=81.163327756 podStartE2EDuration="1m21.163327756s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.162915995 +0000 UTC m=+104.403674514" watchObservedRunningTime="2025-12-09 11:32:59.163327756 +0000 UTC m=+104.404086275" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.175533 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hjqqb" podStartSLOduration=80.175514708 podStartE2EDuration="1m20.175514708s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.175436096 +0000 UTC m=+104.416194615" watchObservedRunningTime="2025-12-09 11:32:59.175514708 +0000 UTC m=+104.416273227" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.193626 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef486a64-888e-4431-ba03-b833fb6e2fca-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.193942 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef486a64-888e-4431-ba03-b833fb6e2fca-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.194116 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.194214 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ef486a64-888e-4431-ba03-b833fb6e2fca-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.194327 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.203051 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.203028764 podStartE2EDuration="21.203028764s" podCreationTimestamp="2025-12-09 11:32:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.200153381 +0000 UTC m=+104.440911900" watchObservedRunningTime="2025-12-09 11:32:59.203028764 +0000 UTC m=+104.443787283" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.258046 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vdpdw" podStartSLOduration=82.258027596 podStartE2EDuration="1m22.258027596s" podCreationTimestamp="2025-12-09 11:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.237814507 +0000 UTC m=+104.478573026" watchObservedRunningTime="2025-12-09 11:32:59.258027596 +0000 UTC m=+104.498786115" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.295195 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef486a64-888e-4431-ba03-b833fb6e2fca-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.295451 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef486a64-888e-4431-ba03-b833fb6e2fca-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.295610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.295701 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.295716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ef486a64-888e-4431-ba03-b833fb6e2fca-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.296122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.296186 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ef486a64-888e-4431-ba03-b833fb6e2fca-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.296942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ef486a64-888e-4431-ba03-b833fb6e2fca-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.298296 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.298277359 podStartE2EDuration="56.298277359s" podCreationTimestamp="2025-12-09 11:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.279793574 +0000 UTC m=+104.520552103" watchObservedRunningTime="2025-12-09 11:32:59.298277359 +0000 UTC m=+104.539035868" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.301640 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef486a64-888e-4431-ba03-b833fb6e2fca-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.314569 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef486a64-888e-4431-ba03-b833fb6e2fca-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr75b\" (UID: \"ef486a64-888e-4431-ba03-b833fb6e2fca\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.339056 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:32:59 crc kubenswrapper[4770]: E1209 11:32:59.339189 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.339425 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:32:59 crc kubenswrapper[4770]: E1209 11:32:59.339498 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.339659 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:32:59 crc kubenswrapper[4770]: E1209 11:32:59.339727 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.339930 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:32:59 crc kubenswrapper[4770]: E1209 11:32:59.339999 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.351338 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podStartSLOduration=81.35131999 podStartE2EDuration="1m21.35131999s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.34975963 +0000 UTC m=+104.590518149" watchObservedRunningTime="2025-12-09 11:32:59.35131999 +0000 UTC m=+104.592078509" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.365689 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-shdpv" podStartSLOduration=81.365658968 podStartE2EDuration="1m21.365658968s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.365131584 +0000 UTC m=+104.605890133" watchObservedRunningTime="2025-12-09 11:32:59.365658968 +0000 UTC m=+104.606417487" Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.369358 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" Dec 09 11:32:59 crc kubenswrapper[4770]: W1209 11:32:59.389060 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef486a64_888e_4431_ba03_b833fb6e2fca.slice/crio-c9bb3437cc5f662338f8e81ca581cdc6991b8d74ba79cdec7b02261ee799cb7d WatchSource:0}: Error finding container c9bb3437cc5f662338f8e81ca581cdc6991b8d74ba79cdec7b02261ee799cb7d: Status 404 returned error can't find the container with id c9bb3437cc5f662338f8e81ca581cdc6991b8d74ba79cdec7b02261ee799cb7d Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.536875 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" event={"ID":"ef486a64-888e-4431-ba03-b833fb6e2fca","Type":"ContainerStarted","Data":"f4a18245ce5a980e38e0eef37ac1ed55df844ddd4ad22c83f18ed4e9d72afed3"} Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.536941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" event={"ID":"ef486a64-888e-4431-ba03-b833fb6e2fca","Type":"ContainerStarted","Data":"c9bb3437cc5f662338f8e81ca581cdc6991b8d74ba79cdec7b02261ee799cb7d"} Dec 09 11:32:59 crc kubenswrapper[4770]: I1209 11:32:59.548523 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr75b" podStartSLOduration=81.54850539 podStartE2EDuration="1m21.54850539s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:32:59.54812278 +0000 UTC m=+104.788881309" watchObservedRunningTime="2025-12-09 11:32:59.54850539 +0000 UTC m=+104.789263899" Dec 09 11:33:01 crc kubenswrapper[4770]: I1209 11:33:01.339320 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:01 crc kubenswrapper[4770]: E1209 11:33:01.339437 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:01 crc kubenswrapper[4770]: I1209 11:33:01.339461 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:01 crc kubenswrapper[4770]: E1209 11:33:01.339590 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:01 crc kubenswrapper[4770]: I1209 11:33:01.339814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:01 crc kubenswrapper[4770]: E1209 11:33:01.339874 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:01 crc kubenswrapper[4770]: I1209 11:33:01.340271 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:01 crc kubenswrapper[4770]: E1209 11:33:01.340326 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:03 crc kubenswrapper[4770]: I1209 11:33:03.339583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:03 crc kubenswrapper[4770]: I1209 11:33:03.339655 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:03 crc kubenswrapper[4770]: E1209 11:33:03.340076 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:03 crc kubenswrapper[4770]: I1209 11:33:03.339736 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:03 crc kubenswrapper[4770]: E1209 11:33:03.340159 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:03 crc kubenswrapper[4770]: I1209 11:33:03.339748 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:03 crc kubenswrapper[4770]: E1209 11:33:03.340238 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:03 crc kubenswrapper[4770]: E1209 11:33:03.340301 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:04 crc kubenswrapper[4770]: I1209 11:33:04.340714 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:33:04 crc kubenswrapper[4770]: E1209 11:33:04.340932 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:33:05 crc kubenswrapper[4770]: I1209 11:33:05.339970 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:05 crc kubenswrapper[4770]: E1209 11:33:05.340112 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:05 crc kubenswrapper[4770]: I1209 11:33:05.340192 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:05 crc kubenswrapper[4770]: I1209 11:33:05.341363 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:05 crc kubenswrapper[4770]: E1209 11:33:05.341416 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:05 crc kubenswrapper[4770]: I1209 11:33:05.341443 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:05 crc kubenswrapper[4770]: E1209 11:33:05.341877 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:05 crc kubenswrapper[4770]: E1209 11:33:05.342030 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:07 crc kubenswrapper[4770]: I1209 11:33:07.340242 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:07 crc kubenswrapper[4770]: I1209 11:33:07.340335 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:07 crc kubenswrapper[4770]: E1209 11:33:07.340402 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:07 crc kubenswrapper[4770]: E1209 11:33:07.340470 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:07 crc kubenswrapper[4770]: I1209 11:33:07.340696 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:07 crc kubenswrapper[4770]: I1209 11:33:07.340735 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:07 crc kubenswrapper[4770]: E1209 11:33:07.340998 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:07 crc kubenswrapper[4770]: E1209 11:33:07.341048 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:09 crc kubenswrapper[4770]: I1209 11:33:09.339622 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:09 crc kubenswrapper[4770]: I1209 11:33:09.339685 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:09 crc kubenswrapper[4770]: I1209 11:33:09.339749 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:09 crc kubenswrapper[4770]: E1209 11:33:09.339753 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:09 crc kubenswrapper[4770]: E1209 11:33:09.339879 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:09 crc kubenswrapper[4770]: E1209 11:33:09.340001 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:09 crc kubenswrapper[4770]: I1209 11:33:09.341017 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:09 crc kubenswrapper[4770]: E1209 11:33:09.341323 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:11 crc kubenswrapper[4770]: I1209 11:33:11.339278 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:11 crc kubenswrapper[4770]: I1209 11:33:11.339369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:11 crc kubenswrapper[4770]: I1209 11:33:11.339385 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:11 crc kubenswrapper[4770]: I1209 11:33:11.339493 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:11 crc kubenswrapper[4770]: E1209 11:33:11.339498 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:11 crc kubenswrapper[4770]: E1209 11:33:11.339600 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:11 crc kubenswrapper[4770]: E1209 11:33:11.339729 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:11 crc kubenswrapper[4770]: E1209 11:33:11.339828 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.339251 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.339318 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.339332 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.339357 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:13 crc kubenswrapper[4770]: E1209 11:33:13.339414 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:13 crc kubenswrapper[4770]: E1209 11:33:13.339465 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:13 crc kubenswrapper[4770]: E1209 11:33:13.339597 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:13 crc kubenswrapper[4770]: E1209 11:33:13.339575 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.578962 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/1.log" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.579573 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/0.log" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.579630 4770 generic.go:334] "Generic (PLEG): container finished" podID="4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86" containerID="b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59" exitCode=1 Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.579660 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerDied","Data":"b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59"} Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.579695 4770 scope.go:117] "RemoveContainer" containerID="440b7d440cf84edb09c8955fcd06fb8689d4ba00e1347a87d3065c6977e8b111" Dec 09 11:33:13 crc kubenswrapper[4770]: I1209 11:33:13.580121 4770 scope.go:117] "RemoveContainer" containerID="b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59" Dec 09 11:33:13 crc kubenswrapper[4770]: E1209 11:33:13.580307 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-q2lxs_openshift-multus(4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86)\"" pod="openshift-multus/multus-q2lxs" podUID="4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86" Dec 09 11:33:14 crc kubenswrapper[4770]: I1209 11:33:14.584782 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/1.log" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.329144 4770 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 09 11:33:15 crc kubenswrapper[4770]: I1209 11:33:15.339583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:15 crc kubenswrapper[4770]: I1209 11:33:15.341005 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.340987 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:15 crc kubenswrapper[4770]: I1209 11:33:15.341120 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:15 crc kubenswrapper[4770]: I1209 11:33:15.341800 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.341876 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.341994 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.342081 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:15 crc kubenswrapper[4770]: E1209 11:33:15.431999 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:33:17 crc kubenswrapper[4770]: I1209 11:33:17.339872 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:17 crc kubenswrapper[4770]: I1209 11:33:17.340066 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:17 crc kubenswrapper[4770]: E1209 11:33:17.340148 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:17 crc kubenswrapper[4770]: I1209 11:33:17.340193 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:17 crc kubenswrapper[4770]: I1209 11:33:17.340256 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:17 crc kubenswrapper[4770]: E1209 11:33:17.340272 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:17 crc kubenswrapper[4770]: E1209 11:33:17.340351 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:17 crc kubenswrapper[4770]: E1209 11:33:17.340508 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:18 crc kubenswrapper[4770]: I1209 11:33:18.340543 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:33:18 crc kubenswrapper[4770]: E1209 11:33:18.340886 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zt2b2_openshift-ovn-kubernetes(eaa774d4-1a7d-4731-a47b-d1d97c88869e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" Dec 09 11:33:19 crc kubenswrapper[4770]: I1209 11:33:19.340166 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:19 crc kubenswrapper[4770]: E1209 11:33:19.340309 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:19 crc kubenswrapper[4770]: I1209 11:33:19.340333 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:19 crc kubenswrapper[4770]: I1209 11:33:19.340393 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:19 crc kubenswrapper[4770]: I1209 11:33:19.340425 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:19 crc kubenswrapper[4770]: E1209 11:33:19.340564 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:19 crc kubenswrapper[4770]: E1209 11:33:19.340700 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:19 crc kubenswrapper[4770]: E1209 11:33:19.340753 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:20 crc kubenswrapper[4770]: E1209 11:33:20.433267 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:33:21 crc kubenswrapper[4770]: I1209 11:33:21.340005 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:21 crc kubenswrapper[4770]: I1209 11:33:21.340039 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:21 crc kubenswrapper[4770]: I1209 11:33:21.340110 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:21 crc kubenswrapper[4770]: E1209 11:33:21.340143 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:21 crc kubenswrapper[4770]: I1209 11:33:21.340175 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:21 crc kubenswrapper[4770]: E1209 11:33:21.340272 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:21 crc kubenswrapper[4770]: E1209 11:33:21.340348 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:21 crc kubenswrapper[4770]: E1209 11:33:21.340392 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:23 crc kubenswrapper[4770]: I1209 11:33:23.339482 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:23 crc kubenswrapper[4770]: I1209 11:33:23.339512 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:23 crc kubenswrapper[4770]: I1209 11:33:23.339498 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:23 crc kubenswrapper[4770]: I1209 11:33:23.339551 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:23 crc kubenswrapper[4770]: E1209 11:33:23.339646 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:23 crc kubenswrapper[4770]: E1209 11:33:23.339846 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:23 crc kubenswrapper[4770]: E1209 11:33:23.340078 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:23 crc kubenswrapper[4770]: E1209 11:33:23.340131 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:25 crc kubenswrapper[4770]: I1209 11:33:25.339555 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:25 crc kubenswrapper[4770]: I1209 11:33:25.339562 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:25 crc kubenswrapper[4770]: I1209 11:33:25.339581 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:25 crc kubenswrapper[4770]: I1209 11:33:25.339585 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:25 crc kubenswrapper[4770]: I1209 11:33:25.340732 4770 scope.go:117] "RemoveContainer" containerID="b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59" Dec 09 11:33:25 crc kubenswrapper[4770]: E1209 11:33:25.340832 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:25 crc kubenswrapper[4770]: E1209 11:33:25.340948 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:25 crc kubenswrapper[4770]: E1209 11:33:25.341014 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:25 crc kubenswrapper[4770]: E1209 11:33:25.340729 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:25 crc kubenswrapper[4770]: E1209 11:33:25.433679 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:33:26 crc kubenswrapper[4770]: I1209 11:33:26.622094 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/1.log" Dec 09 11:33:26 crc kubenswrapper[4770]: I1209 11:33:26.622154 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerStarted","Data":"55d7d7cac78fc0627d84150710ccaaeb7dba76a2a111e5f3e2b56f76be78c372"} Dec 09 11:33:27 crc kubenswrapper[4770]: I1209 11:33:27.340210 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:27 crc kubenswrapper[4770]: I1209 11:33:27.340265 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:27 crc kubenswrapper[4770]: I1209 11:33:27.340339 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:27 crc kubenswrapper[4770]: I1209 11:33:27.340456 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:27 crc kubenswrapper[4770]: E1209 11:33:27.340625 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:27 crc kubenswrapper[4770]: E1209 11:33:27.340815 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:27 crc kubenswrapper[4770]: E1209 11:33:27.341019 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:27 crc kubenswrapper[4770]: E1209 11:33:27.341112 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.339796 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.339888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.339831 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.339989 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:29 crc kubenswrapper[4770]: E1209 11:33:29.340029 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:29 crc kubenswrapper[4770]: E1209 11:33:29.340184 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:29 crc kubenswrapper[4770]: E1209 11:33:29.340261 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:29 crc kubenswrapper[4770]: E1209 11:33:29.340865 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.341539 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.632257 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.634662 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerStarted","Data":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.635451 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:33:29 crc kubenswrapper[4770]: I1209 11:33:29.661046 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podStartSLOduration=111.661029371 podStartE2EDuration="1m51.661029371s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:29.660142978 +0000 UTC m=+134.900901497" watchObservedRunningTime="2025-12-09 11:33:29.661029371 +0000 UTC m=+134.901787890" Dec 09 11:33:30 crc kubenswrapper[4770]: I1209 11:33:30.141310 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8pmjn"] Dec 09 11:33:30 crc kubenswrapper[4770]: I1209 11:33:30.141445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:30 crc kubenswrapper[4770]: E1209 11:33:30.141555 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:30 crc kubenswrapper[4770]: E1209 11:33:30.435191 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:33:31 crc kubenswrapper[4770]: I1209 11:33:31.339106 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:31 crc kubenswrapper[4770]: I1209 11:33:31.339188 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:31 crc kubenswrapper[4770]: I1209 11:33:31.339185 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:31 crc kubenswrapper[4770]: E1209 11:33:31.339309 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:31 crc kubenswrapper[4770]: I1209 11:33:31.339388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:31 crc kubenswrapper[4770]: E1209 11:33:31.339438 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:31 crc kubenswrapper[4770]: E1209 11:33:31.339399 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:31 crc kubenswrapper[4770]: E1209 11:33:31.339548 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:33 crc kubenswrapper[4770]: I1209 11:33:33.339759 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:33 crc kubenswrapper[4770]: I1209 11:33:33.339849 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:33 crc kubenswrapper[4770]: I1209 11:33:33.339783 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:33 crc kubenswrapper[4770]: I1209 11:33:33.339977 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:33 crc kubenswrapper[4770]: E1209 11:33:33.340059 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:33 crc kubenswrapper[4770]: E1209 11:33:33.340244 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:33 crc kubenswrapper[4770]: E1209 11:33:33.340429 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:33 crc kubenswrapper[4770]: E1209 11:33:33.340541 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:35 crc kubenswrapper[4770]: I1209 11:33:35.339957 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:35 crc kubenswrapper[4770]: I1209 11:33:35.340030 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:35 crc kubenswrapper[4770]: I1209 11:33:35.340063 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:35 crc kubenswrapper[4770]: I1209 11:33:35.340030 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:35 crc kubenswrapper[4770]: E1209 11:33:35.341088 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8pmjn" podUID="95971ae2-d7fe-4259-b041-9b6373d8a51c" Dec 09 11:33:35 crc kubenswrapper[4770]: E1209 11:33:35.341186 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 11:33:35 crc kubenswrapper[4770]: E1209 11:33:35.341268 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 11:33:35 crc kubenswrapper[4770]: E1209 11:33:35.341334 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.339081 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.339081 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.339120 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.339336 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343178 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343239 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343190 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343461 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343584 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 11:33:37 crc kubenswrapper[4770]: I1209 11:33:37.343677 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.599314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.647498 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.648061 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.649519 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-j4njp"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.650040 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.650203 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5nk5g"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.650712 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.651730 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.652044 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.652096 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.652211 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.653639 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.658143 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.658711 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.659151 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-42jsz"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.659477 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.659531 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.659698 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.660261 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661175 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661349 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661396 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661557 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661654 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661794 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.661840 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662077 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662198 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662262 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662098 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662401 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662528 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662615 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.662546 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.664450 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.666016 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.666231 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.667745 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.669446 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.670041 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.670488 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.670723 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.670970 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.671142 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.671773 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wdxh"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.672001 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.672158 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.672233 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.674744 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xwjv4"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.675283 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.676427 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.676659 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.677377 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.677609 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.678270 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.678658 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.678712 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.680344 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44f6s"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.680404 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.681017 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.681819 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.681880 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.682916 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.683543 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.684414 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.684612 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.684795 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.684994 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.685638 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.685821 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.685857 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.685986 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.685997 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686101 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686138 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686235 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686247 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686274 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686287 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686239 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.684618 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686443 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686276 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686527 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.686567 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.707825 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.708752 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.709634 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.710296 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.710303 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.715771 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.715941 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.726162 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.726718 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.729913 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.732596 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-j4njp"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.732684 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.732875 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733598 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv8dq\" (UniqueName: \"kubernetes.io/projected/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-kube-api-access-zv8dq\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733669 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75426\" (UniqueName: \"kubernetes.io/projected/7ed604a1-90f7-4f46-a4a3-8e097b02c922-kube-api-access-75426\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-encryption-config\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733713 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733738 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733756 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-node-pullsecrets\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733801 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-config\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733820 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c86e7caf-f2ca-4006-a807-e055f0597bf9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733843 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phg52\" (UniqueName: \"kubernetes.io/projected/25c0c5e8-78eb-427b-88f9-e02365a6afc1-kube-api-access-phg52\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733862 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733883 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733926 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z24wg\" (UniqueName: \"kubernetes.io/projected/43156f0f-537c-4483-827e-d02941e9a274-kube-api-access-z24wg\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733964 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-encryption-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.733986 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734019 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgzx\" (UniqueName: \"kubernetes.io/projected/e69f8571-c321-4ff0-9fd1-fb67b1176230-kube-api-access-qxgzx\") pod \"downloads-7954f5f757-xwjv4\" (UID: \"e69f8571-c321-4ff0-9fd1-fb67b1176230\") " pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734042 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734081 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734108 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734131 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-serving-cert\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734163 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjblr\" (UniqueName: \"kubernetes.io/projected/c891628b-1ab0-498d-8792-60029bbff6b3-kube-api-access-hjblr\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734181 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734200 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmp5h\" (UniqueName: \"kubernetes.io/projected/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-kube-api-access-bmp5h\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734225 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734254 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-client\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734274 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734291 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734311 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27k9\" (UniqueName: \"kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734328 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-client\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734347 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-dir\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734368 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58j25\" (UniqueName: \"kubernetes.io/projected/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-kube-api-access-58j25\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734387 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734410 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0c869f7-6828-4624-be3c-1449ffc2e514-machine-approver-tls\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734429 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc528\" (UniqueName: \"kubernetes.io/projected/d0c869f7-6828-4624-be3c-1449ffc2e514-kube-api-access-xc528\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734463 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734484 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734503 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dj9n\" (UniqueName: \"kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734523 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-config\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734546 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734568 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c891628b-1ab0-498d-8792-60029bbff6b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734623 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-trusted-ca\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734643 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734665 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734704 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734727 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dlc6\" (UniqueName: \"kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ed604a1-90f7-4f46-a4a3-8e097b02c922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734794 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86e7caf-f2ca-4006-a807-e055f0597bf9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734813 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734837 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734855 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-serving-cert\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734890 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734926 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734951 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-auth-proxy-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734977 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.734996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-image-import-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735017 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735035 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735056 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735074 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmbt\" (UniqueName: \"kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735096 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c891628b-1ab0-498d-8792-60029bbff6b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-config\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735163 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-policies\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735210 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnt4c\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-kube-api-access-mnt4c\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735231 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735248 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit-dir\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735268 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-service-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735291 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rjm8\" (UniqueName: \"kubernetes.io/projected/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-kube-api-access-5rjm8\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735311 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735343 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-images\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735366 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43156f0f-537c-4483-827e-d02941e9a274-serving-cert\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735385 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-serving-cert\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.735404 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.737263 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738389 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738472 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738655 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738764 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738798 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.738832 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.740088 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.740195 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.740282 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.740988 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.741648 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742177 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742454 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742511 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742583 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742693 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742836 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.742986 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743008 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743155 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743270 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743473 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743560 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5nk5g"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.743573 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.744376 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.744413 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.745135 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.750408 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.755423 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.756413 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44f6s"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.760534 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.761783 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xwjv4"] Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.919508 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.919643 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920088 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920112 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-serving-cert\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920481 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv8dq\" (UniqueName: \"kubernetes.io/projected/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-kube-api-access-zv8dq\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920517 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75426\" (UniqueName: \"kubernetes.io/projected/7ed604a1-90f7-4f46-a4a3-8e097b02c922-kube-api-access-75426\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920535 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-encryption-config\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920568 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:39 crc kubenswrapper[4770]: I1209 11:33:39.920584 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-node-pullsecrets\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.880771 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.880861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-config\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.880918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c86e7caf-f2ca-4006-a807-e055f0597bf9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.880959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.880998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phg52\" (UniqueName: \"kubernetes.io/projected/25c0c5e8-78eb-427b-88f9-e02365a6afc1-kube-api-access-phg52\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881028 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881056 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881085 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881142 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z24wg\" (UniqueName: \"kubernetes.io/projected/43156f0f-537c-4483-827e-d02941e9a274-kube-api-access-z24wg\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-encryption-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881200 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881230 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881282 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgzx\" (UniqueName: \"kubernetes.io/projected/e69f8571-c321-4ff0-9fd1-fb67b1176230-kube-api-access-qxgzx\") pod \"downloads-7954f5f757-xwjv4\" (UID: \"e69f8571-c321-4ff0-9fd1-fb67b1176230\") " pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881305 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881333 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881366 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-serving-cert\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881405 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881460 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjblr\" (UniqueName: \"kubernetes.io/projected/c891628b-1ab0-498d-8792-60029bbff6b3-kube-api-access-hjblr\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881499 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmp5h\" (UniqueName: \"kubernetes.io/projected/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-kube-api-access-bmp5h\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881531 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881559 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-client\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881588 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27k9\" (UniqueName: \"kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881614 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-client\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881641 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-dir\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881664 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881684 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881725 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0c869f7-6828-4624-be3c-1449ffc2e514-machine-approver-tls\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881763 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc528\" (UniqueName: \"kubernetes.io/projected/d0c869f7-6828-4624-be3c-1449ffc2e514-kube-api-access-xc528\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58j25\" (UniqueName: \"kubernetes.io/projected/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-kube-api-access-58j25\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881819 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881854 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881938 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dj9n\" (UniqueName: \"kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-config\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.881983 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882006 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882043 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c891628b-1ab0-498d-8792-60029bbff6b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-trusted-ca\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882094 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882188 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ed604a1-90f7-4f46-a4a3-8e097b02c922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882213 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86e7caf-f2ca-4006-a807-e055f0597bf9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882315 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882345 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882389 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dlc6\" (UniqueName: \"kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882418 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882448 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-serving-cert\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882575 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882604 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882631 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-image-import-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-auth-proxy-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882933 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.882998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883109 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmbt\" (UniqueName: \"kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-config\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883275 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-policies\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c891628b-1ab0-498d-8792-60029bbff6b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883339 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnt4c\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-kube-api-access-mnt4c\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883360 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit-dir\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883441 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883460 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-service-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883544 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rjm8\" (UniqueName: \"kubernetes.io/projected/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-kube-api-access-5rjm8\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883565 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883619 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-images\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.883693 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43156f0f-537c-4483-827e-d02941e9a274-serving-cert\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.896002 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.896209 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.896583 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.896982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.897483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-node-pullsecrets\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.898603 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.898644 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43156f0f-537c-4483-827e-d02941e9a274-serving-cert\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.900443 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.910483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-encryption-config\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.915523 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c86e7caf-f2ca-4006-a807-e055f0597bf9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.916315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.918377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-serving-cert\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.918498 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-image-import-ca\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.918705 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.918734 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.918977 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-serving-cert\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.919862 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wdxh"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.923195 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.901298 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-config\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.924499 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.924540 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.924640 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.927526 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25c0c5e8-78eb-427b-88f9-e02365a6afc1-etcd-client\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.902036 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c891628b-1ab0-498d-8792-60029bbff6b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.929750 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.930073 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-config\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.930137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.931530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0c869f7-6828-4624-be3c-1449ffc2e514-auth-proxy-config\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.932569 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86e7caf-f2ca-4006-a807-e055f0597bf9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.933079 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.933445 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.934475 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-trusted-ca\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.935606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c891628b-1ab0-498d-8792-60029bbff6b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.936094 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.936170 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.936220 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.936216 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937352 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937600 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-config\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937682 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ed604a1-90f7-4f46-a4a3-8e097b02c922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937774 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit-dir\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.937989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.938205 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.938693 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-policies\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.939084 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.939120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.939390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-service-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.939442 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25c0c5e8-78eb-427b-88f9-e02365a6afc1-audit-dir\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.939598 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.940540 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.941090 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.941460 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43156f0f-537c-4483-827e-d02941e9a274-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.943818 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7ed604a1-90f7-4f46-a4a3-8e097b02c922-images\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.946490 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.947000 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-encryption-config\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.947467 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-audit\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.949267 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.949546 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.953369 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0c869f7-6828-4624-be3c-1449ffc2e514-machine-approver-tls\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.954120 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.957865 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-42jsz"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.958029 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.963139 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.963743 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.965111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.965999 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.966675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.967198 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8t95q"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.968618 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.972034 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-serving-cert\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.973219 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.973573 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-etcd-client\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.973686 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.973712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.973945 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.975236 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.975462 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.975768 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.976225 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.976335 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.976424 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.976481 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.978574 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.982465 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z24wg\" (UniqueName: \"kubernetes.io/projected/43156f0f-537c-4483-827e-d02941e9a274-kube-api-access-z24wg\") pod \"authentication-operator-69f744f599-42jsz\" (UID: \"43156f0f-537c-4483-827e-d02941e9a274\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.983166 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.983291 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.983389 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.983754 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.984127 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.984725 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.984814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.985317 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv8dq\" (UniqueName: \"kubernetes.io/projected/c0d5b7a7-6e5f-430e-863a-a18ac009ee3c-kube-api-access-zv8dq\") pod \"openshift-apiserver-operator-796bbdcf4f-b6p2f\" (UID: \"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.986037 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.986883 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.987350 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.987816 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.988315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dj9n\" (UniqueName: \"kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n\") pod \"route-controller-manager-6576b87f9c-d8ftw\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.990947 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.990985 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.991372 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.991380 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.991520 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.992245 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.992276 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.992302 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.992774 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.993147 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjblr\" (UniqueName: \"kubernetes.io/projected/c891628b-1ab0-498d-8792-60029bbff6b3-kube-api-access-hjblr\") pod \"openshift-config-operator-7777fb866f-wdt8r\" (UID: \"c891628b-1ab0-498d-8792-60029bbff6b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.993306 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75426\" (UniqueName: \"kubernetes.io/projected/7ed604a1-90f7-4f46-a4a3-8e097b02c922-kube-api-access-75426\") pod \"machine-api-operator-5694c8668f-5nk5g\" (UID: \"7ed604a1-90f7-4f46-a4a3-8e097b02c922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.995538 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.995786 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.996758 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.996786 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.997279 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phg52\" (UniqueName: \"kubernetes.io/projected/25c0c5e8-78eb-427b-88f9-e02365a6afc1-kube-api-access-phg52\") pod \"apiserver-7bbb656c7d-dzk4z\" (UID: \"25c0c5e8-78eb-427b-88f9-e02365a6afc1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.997607 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmbt\" (UniqueName: \"kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.997641 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk"] Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.998111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.998601 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:40 crc kubenswrapper[4770]: I1209 11:33:40.999117 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.000557 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.001802 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.002996 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgzx\" (UniqueName: \"kubernetes.io/projected/e69f8571-c321-4ff0-9fd1-fb67b1176230-kube-api-access-qxgzx\") pod \"downloads-7954f5f757-xwjv4\" (UID: \"e69f8571-c321-4ff0-9fd1-fb67b1176230\") " pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.003674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27k9\" (UniqueName: \"kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9\") pod \"controller-manager-879f6c89f-frwww\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.003726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnt4c\" (UniqueName: \"kubernetes.io/projected/c86e7caf-f2ca-4006-a807-e055f0597bf9-kube-api-access-mnt4c\") pod \"cluster-image-registry-operator-dc59b4c8b-qr7cm\" (UID: \"c86e7caf-f2ca-4006-a807-e055f0597bf9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.003887 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.004065 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.004592 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.005428 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rjm8\" (UniqueName: \"kubernetes.io/projected/e215f2a5-db5f-4828-8824-bd30bc0f6c4d-kube-api-access-5rjm8\") pod \"apiserver-76f77b778f-j4njp\" (UID: \"e215f2a5-db5f-4828-8824-bd30bc0f6c4d\") " pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.004806 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.006555 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wdxh\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.006862 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc528\" (UniqueName: \"kubernetes.io/projected/d0c869f7-6828-4624-be3c-1449ffc2e514-kube-api-access-xc528\") pod \"machine-approver-56656f9798-cv4wq\" (UID: \"d0c869f7-6828-4624-be3c-1449ffc2e514\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.011044 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58j25\" (UniqueName: \"kubernetes.io/projected/ae983b39-dbad-4db0-97f9-2bab5bb3ef36-kube-api-access-58j25\") pod \"console-operator-58897d9998-44f6s\" (UID: \"ae983b39-dbad-4db0-97f9-2bab5bb3ef36\") " pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.011879 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dlc6\" (UniqueName: \"kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6\") pod \"console-f9d7485db-mhgdc\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.011953 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.012690 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.012804 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fsglb"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.015101 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmp5h\" (UniqueName: \"kubernetes.io/projected/45e8e8a9-2925-41c1-b8f5-7aafde0d700d-kube-api-access-bmp5h\") pod \"cluster-samples-operator-665b6dd947-dg9hr\" (UID: \"45e8e8a9-2925-41c1-b8f5-7aafde0d700d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.017566 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.017673 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.018001 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.018191 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.018362 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.019469 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9tllw"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.022814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.024197 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.032217 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.036220 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.042151 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.043017 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.044263 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.045477 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.046636 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.046700 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.046647 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.046892 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.051024 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.052569 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.053053 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.053524 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.054247 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.054522 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.054594 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.057392 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9xgcj"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.057774 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.057873 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.058149 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.058337 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.063632 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.067448 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.068228 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-64jcp"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.068734 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.068763 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.070316 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.071525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.072977 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.073536 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.074147 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.074194 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.075628 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.076952 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.077384 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.085354 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-67vrh"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.086121 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.086624 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwlr8"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.087097 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.087456 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088219 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kvqt\" (UniqueName: \"kubernetes.io/projected/8e1f7c2c-c165-409d-9e75-a22656ae8c12-kube-api-access-7kvqt\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088368 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a56c07-a692-4268-b1d1-e7535e738ad3-proxy-tls\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088440 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9qht\" (UniqueName: \"kubernetes.io/projected/76a56c07-a692-4268-b1d1-e7535e738ad3-kube-api-access-p9qht\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088459 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088479 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32fdb059-5acb-446d-98ac-8239919a90b7-metrics-tls\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25fkc\" (UniqueName: \"kubernetes.io/projected/32fdb059-5acb-446d-98ac-8239919a90b7-kube-api-access-25fkc\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088543 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrjs2\" (UniqueName: \"kubernetes.io/projected/a58a8c2f-222c-41c7-8189-2610471f69df-kube-api-access-jrjs2\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088572 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-metrics-tls\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088596 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n68w\" (UniqueName: \"kubernetes.io/projected/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-kube-api-access-5n68w\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088638 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-serving-cert\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088669 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-trusted-ca\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1f7c2c-c165-409d-9e75-a22656ae8c12-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088785 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-srv-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088811 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktnk\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-kube-api-access-fktnk\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-config\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088873 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz5kw\" (UniqueName: \"kubernetes.io/projected/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-kube-api-access-dz5kw\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.088998 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjpx2\" (UniqueName: \"kubernetes.io/projected/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-kube-api-access-mjpx2\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.089068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e1f7c2c-c165-409d-9e75-a22656ae8c12-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.089093 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-images\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.093926 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.094274 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vpb8g"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.095433 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.095895 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.096701 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.102103 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-q9hbn"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.102363 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.102998 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ftbnh"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103410 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8t95q"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103425 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103435 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9tllw"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103444 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103454 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103462 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103472 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9xgcj"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103540 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.103748 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.104744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.106422 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fsglb"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.107759 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.109096 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-64jcp"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.110975 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.115556 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.115623 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.116976 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q9hbn"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.118099 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.120008 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.120519 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vpb8g"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.122833 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.124077 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.125261 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.126602 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.145147 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.145215 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.145230 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwlr8"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.150308 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.150565 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.151273 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.158134 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c"] Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.174580 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189616 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-metrics-tls\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n68w\" (UniqueName: \"kubernetes.io/projected/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-kube-api-access-5n68w\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189666 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-serving-cert\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189685 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-trusted-ca\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1f7c2c-c165-409d-9e75-a22656ae8c12-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189746 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-srv-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189762 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktnk\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-kube-api-access-fktnk\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-config\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189792 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz5kw\" (UniqueName: \"kubernetes.io/projected/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-kube-api-access-dz5kw\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189831 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjpx2\" (UniqueName: \"kubernetes.io/projected/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-kube-api-access-mjpx2\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e1f7c2c-c165-409d-9e75-a22656ae8c12-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189873 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-images\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189916 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189953 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kvqt\" (UniqueName: \"kubernetes.io/projected/8e1f7c2c-c165-409d-9e75-a22656ae8c12-kube-api-access-7kvqt\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.189982 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190022 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a56c07-a692-4268-b1d1-e7535e738ad3-proxy-tls\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190051 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190071 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9qht\" (UniqueName: \"kubernetes.io/projected/76a56c07-a692-4268-b1d1-e7535e738ad3-kube-api-access-p9qht\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190095 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190118 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32fdb059-5acb-446d-98ac-8239919a90b7-metrics-tls\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrjs2\" (UniqueName: \"kubernetes.io/projected/a58a8c2f-222c-41c7-8189-2610471f69df-kube-api-access-jrjs2\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.190162 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25fkc\" (UniqueName: \"kubernetes.io/projected/32fdb059-5acb-446d-98ac-8239919a90b7-kube-api-access-25fkc\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.196653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1f7c2c-c165-409d-9e75-a22656ae8c12-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.197392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.198238 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-trusted-ca\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.199365 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-srv-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.200083 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-config\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.201141 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.212703 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.247411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.247482 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.249112 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.249480 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.250505 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e1f7c2c-c165-409d-9e75-a22656ae8c12-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.251573 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.253539 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.255385 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.256248 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.265970 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.273066 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.274972 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.278194 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a56c07-a692-4268-b1d1-e7535e738ad3-proxy-tls\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.278463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-images\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.278463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a56c07-a692-4268-b1d1-e7535e738ad3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.279124 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-metrics-tls\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.279228 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-serving-cert\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.282392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a58a8c2f-222c-41c7-8189-2610471f69df-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.282477 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32fdb059-5acb-446d-98ac-8239919a90b7-metrics-tls\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.291030 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.314335 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.361301 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.361712 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.371939 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.392342 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.392578 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.392931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.393009 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.393892 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:35:43.393861534 +0000 UTC m=+268.634620103 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.394940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.403485 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.410975 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.461005 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.501673 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.501754 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.504290 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.505592 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.508172 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.508180 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.511706 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.513325 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.593071 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.593627 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.593978 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.600202 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.600455 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.600668 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607658 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607739 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lww2l\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607834 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607858 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607887 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.607962 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.608095 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.611699 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.111681941 +0000 UTC m=+147.352440450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.673602 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.675418 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.676076 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.677729 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.695285 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.709815 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.709884 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj7sg\" (UniqueName: \"kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.709951 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.709968 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-metrics-certs\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.709987 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lww2l\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710003 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710021 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710038 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-registration-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710087 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqhdp\" (UniqueName: \"kubernetes.io/projected/4d96f97a-35bd-4d9f-998b-0b1316c5e39d-kube-api-access-lqhdp\") pod \"migrator-59844c95c7-8hkwp\" (UID: \"4d96f97a-35bd-4d9f-998b-0b1316c5e39d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710101 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-config\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710126 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710148 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710162 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-cert\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710177 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710193 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7df0970-4671-4495-8c49-b1c2c096885e-signing-key\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710206 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-node-bootstrap-token\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710222 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pknf9\" (UniqueName: \"kubernetes.io/projected/25b57e6c-de38-4c55-9960-b86dccd6b3e4-kube-api-access-pknf9\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710247 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-apiservice-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710262 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2mhm\" (UniqueName: \"kubernetes.io/projected/f0fc13ea-940c-43f5-b256-245017e6664f-kube-api-access-v2mhm\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710283 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7df0970-4671-4495-8c49-b1c2c096885e-signing-cabundle\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710297 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnljg\" (UniqueName: \"kubernetes.io/projected/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-kube-api-access-wnljg\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710335 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvf5s\" (UniqueName: \"kubernetes.io/projected/365d6032-3c70-47f9-ae54-632097bd0a47-kube-api-access-xvf5s\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710349 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f021879-eaf1-4d8b-8d8f-bd91525114c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710364 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-client\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710378 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/365d6032-3c70-47f9-ae54-632097bd0a47-proxy-tls\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710412 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-stats-auth\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710426 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-socket-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710441 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wrnx\" (UniqueName: \"kubernetes.io/projected/6d970f00-0a9d-4259-b21a-503e1e0bceb2-kube-api-access-9wrnx\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710458 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-config\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710491 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/365d6032-3c70-47f9-ae54-632097bd0a47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd1aa89d-775c-47be-a2ba-bf592371938d-metrics-tls\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710623 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710647 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.710689 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-webhook-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716149 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-service-ca-bundle\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716246 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-plugins-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716321 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1aa89d-775c-47be-a2ba-bf592371938d-config-volume\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716404 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6sgd\" (UniqueName: \"kubernetes.io/projected/a7df0970-4671-4495-8c49-b1c2c096885e-kube-api-access-g6sgd\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716470 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq6k9\" (UniqueName: \"kubernetes.io/projected/3dc28a15-7069-4b05-aced-372d68a397ff-kube-api-access-rq6k9\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716540 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9gvr\" (UniqueName: \"kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716605 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-certs\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716671 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-service-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716738 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cnbd\" (UniqueName: \"kubernetes.io/projected/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-kube-api-access-5cnbd\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716805 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716867 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.716956 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-mountpoint-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717026 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-csi-data-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717099 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717169 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f021879-eaf1-4d8b-8d8f-bd91525114c3-config\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717241 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-srv-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717428 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmkn5\" (UniqueName: \"kubernetes.io/projected/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-kube-api-access-bmkn5\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.717536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6d970f00-0a9d-4259-b21a-503e1e0bceb2-tmpfs\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.718819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.718928 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl9x7\" (UniqueName: \"kubernetes.io/projected/bd1aa89d-775c-47be-a2ba-bf592371938d-kube-api-access-pl9x7\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g242h\" (UniqueName: \"kubernetes.io/projected/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-kube-api-access-g242h\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720356 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f021879-eaf1-4d8b-8d8f-bd91525114c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720425 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-default-certificate\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720556 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720715 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-serving-cert\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.720717 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.721062 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.719595 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.722142 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.221378614 +0000 UTC m=+147.462137133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.723853 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.723963 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq8x9\" (UniqueName: \"kubernetes.io/projected/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-kube-api-access-vq8x9\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.728409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.730503 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.751513 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.769917 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.794840 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.795151 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.797530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.812741 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.828862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj7sg\" (UniqueName: \"kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.828922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.828943 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.828961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-metrics-certs\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.828982 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829003 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829022 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-registration-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829039 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829057 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqhdp\" (UniqueName: \"kubernetes.io/projected/4d96f97a-35bd-4d9f-998b-0b1316c5e39d-kube-api-access-lqhdp\") pod \"migrator-59844c95c7-8hkwp\" (UID: \"4d96f97a-35bd-4d9f-998b-0b1316c5e39d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829072 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-config\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829091 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-cert\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829139 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7df0970-4671-4495-8c49-b1c2c096885e-signing-key\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829174 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-node-bootstrap-token\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829189 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pknf9\" (UniqueName: \"kubernetes.io/projected/25b57e6c-de38-4c55-9960-b86dccd6b3e4-kube-api-access-pknf9\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829227 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-apiservice-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2mhm\" (UniqueName: \"kubernetes.io/projected/f0fc13ea-940c-43f5-b256-245017e6664f-kube-api-access-v2mhm\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829282 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7df0970-4671-4495-8c49-b1c2c096885e-signing-cabundle\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829303 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnljg\" (UniqueName: \"kubernetes.io/projected/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-kube-api-access-wnljg\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829327 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829355 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvf5s\" (UniqueName: \"kubernetes.io/projected/365d6032-3c70-47f9-ae54-632097bd0a47-kube-api-access-xvf5s\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f021879-eaf1-4d8b-8d8f-bd91525114c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829410 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-client\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829447 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/365d6032-3c70-47f9-ae54-632097bd0a47-proxy-tls\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-stats-auth\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829478 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-socket-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829495 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wrnx\" (UniqueName: \"kubernetes.io/projected/6d970f00-0a9d-4259-b21a-503e1e0bceb2-kube-api-access-9wrnx\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829525 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-config\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829548 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829574 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/365d6032-3c70-47f9-ae54-632097bd0a47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829601 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd1aa89d-775c-47be-a2ba-bf592371938d-metrics-tls\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829682 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829701 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-webhook-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-service-ca-bundle\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-plugins-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1aa89d-775c-47be-a2ba-bf592371938d-config-volume\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829842 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6sgd\" (UniqueName: \"kubernetes.io/projected/a7df0970-4671-4495-8c49-b1c2c096885e-kube-api-access-g6sgd\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829860 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq6k9\" (UniqueName: \"kubernetes.io/projected/3dc28a15-7069-4b05-aced-372d68a397ff-kube-api-access-rq6k9\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9gvr\" (UniqueName: \"kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-certs\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829966 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-service-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.829985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cnbd\" (UniqueName: \"kubernetes.io/projected/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-kube-api-access-5cnbd\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830013 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-mountpoint-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830049 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-csi-data-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f021879-eaf1-4d8b-8d8f-bd91525114c3-config\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830100 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-srv-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830136 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmkn5\" (UniqueName: \"kubernetes.io/projected/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-kube-api-access-bmkn5\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6d970f00-0a9d-4259-b21a-503e1e0bceb2-tmpfs\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830188 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830217 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl9x7\" (UniqueName: \"kubernetes.io/projected/bd1aa89d-775c-47be-a2ba-bf592371938d-kube-api-access-pl9x7\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830239 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g242h\" (UniqueName: \"kubernetes.io/projected/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-kube-api-access-g242h\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830254 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f021879-eaf1-4d8b-8d8f-bd91525114c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830273 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-default-certificate\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830301 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.830335 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.330310828 +0000 UTC m=+147.571069347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-serving-cert\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830453 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq8x9\" (UniqueName: \"kubernetes.io/projected/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-kube-api-access-vq8x9\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.830884 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.831184 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-registration-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.833519 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.835734 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.836412 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7df0970-4671-4495-8c49-b1c2c096885e-signing-cabundle\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.836889 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.837652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7df0970-4671-4495-8c49-b1c2c096885e-signing-key\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.838454 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-config\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.839895 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-client\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.844096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-mountpoint-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.844364 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-csi-data-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.845125 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f021879-eaf1-4d8b-8d8f-bd91525114c3-config\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.845462 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-socket-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.846501 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-config\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.847339 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/365d6032-3c70-47f9-ae54-632097bd0a47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.847938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6d970f00-0a9d-4259-b21a-503e1e0bceb2-tmpfs\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.849762 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.851420 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.852090 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1aa89d-775c-47be-a2ba-bf592371938d-config-volume\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.852174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-plugins-dir\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.852911 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd1aa89d-775c-47be-a2ba-bf592371938d-metrics-tls\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.854171 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.854784 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.855824 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.856575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-webhook-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.857480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.862875 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3dc28a15-7069-4b05-aced-372d68a397ff-srv-cert\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.863642 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.864298 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-certs\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.865278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/365d6032-3c70-47f9-ae54-632097bd0a47-proxy-tls\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.865756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.867741 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0fc13ea-940c-43f5-b256-245017e6664f-serving-cert\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.871156 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.872213 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f021879-eaf1-4d8b-8d8f-bd91525114c3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.873388 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.875038 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d970f00-0a9d-4259-b21a-503e1e0bceb2-apiservice-cert\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.876470 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0fc13ea-940c-43f5-b256-245017e6664f-etcd-service-ca\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.880360 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/25b57e6c-de38-4c55-9960-b86dccd6b3e4-node-bootstrap-token\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.934548 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.934794 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.434756617 +0000 UTC m=+147.675515136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.935689 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.936004 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.944114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.946077 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 11:33:41 crc kubenswrapper[4770]: E1209 11:33:41.948740 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.448723505 +0000 UTC m=+147.689482024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.949275 4770 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 11:33:41 crc kubenswrapper[4770]: I1209 11:33:41.970249 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:41.989811 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.044579 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.044656 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.046275 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.046629 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.546605006 +0000 UTC m=+147.787363525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.047040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.047997 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.547977741 +0000 UTC m=+147.788736260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.056155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" event={"ID":"d0c869f7-6828-4624-be3c-1449ffc2e514","Type":"ContainerStarted","Data":"2e07027deb3b01d8edb1b3ca7560fc4958e2956c778197e0c14442cffc2a6ebe"} Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.065297 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-stats-auth\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.065420 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.069790 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.115433 4770 request.go:700] Waited for 1.011331459s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.118690 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.118861 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.125677 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-default-certificate\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.130529 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-metrics-certs\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.144529 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.146335 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-service-ca-bundle\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.148644 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.149336 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.64931021 +0000 UTC m=+147.890068719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.170744 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.171173 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.207666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-cert\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.213824 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.278265 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.278729 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.778716229 +0000 UTC m=+148.019474748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.287888 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25fkc\" (UniqueName: \"kubernetes.io/projected/32fdb059-5acb-446d-98ac-8239919a90b7-kube-api-access-25fkc\") pod \"dns-operator-744455d44c-8t95q\" (UID: \"32fdb059-5acb-446d-98ac-8239919a90b7\") " pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.305884 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjpx2\" (UniqueName: \"kubernetes.io/projected/4e08edb8-6c28-49a2-ba80-7e3703f48f5f-kube-api-access-mjpx2\") pod \"control-plane-machine-set-operator-78cbb6b69f-bhktd\" (UID: \"4e08edb8-6c28-49a2-ba80-7e3703f48f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.395762 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.397125 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.398210 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.898185513 +0000 UTC m=+148.138944032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.452047 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.462606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz5kw\" (UniqueName: \"kubernetes.io/projected/6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77-kube-api-access-dz5kw\") pod \"openshift-controller-manager-operator-756b6f6bc6-89tfr\" (UID: \"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.466206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktnk\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-kube-api-access-fktnk\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.478680 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/412ed7a7-4ea3-4daf-8978-a70cf1702c1e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2gwrb\" (UID: \"412ed7a7-4ea3-4daf-8978-a70cf1702c1e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.485463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n68w\" (UniqueName: \"kubernetes.io/projected/e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9-kube-api-access-5n68w\") pod \"service-ca-operator-777779d784-fsglb\" (UID: \"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.485850 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f021879-eaf1-4d8b-8d8f-bd91525114c3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ppj6k\" (UID: \"1f021879-eaf1-4d8b-8d8f-bd91525114c3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.486071 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrjs2\" (UniqueName: \"kubernetes.io/projected/a58a8c2f-222c-41c7-8189-2610471f69df-kube-api-access-jrjs2\") pod \"olm-operator-6b444d44fb-cz9jk\" (UID: \"a58a8c2f-222c-41c7-8189-2610471f69df\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.486596 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.486669 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9qht\" (UniqueName: \"kubernetes.io/projected/76a56c07-a692-4268-b1d1-e7535e738ad3-kube-api-access-p9qht\") pod \"machine-config-operator-74547568cd-5vbfw\" (UID: \"76a56c07-a692-4268-b1d1-e7535e738ad3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.487292 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kvqt\" (UniqueName: \"kubernetes.io/projected/8e1f7c2c-c165-409d-9e75-a22656ae8c12-kube-api-access-7kvqt\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlkvz\" (UID: \"8e1f7c2c-c165-409d-9e75-a22656ae8c12\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.492347 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj7sg\" (UniqueName: \"kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg\") pod \"collect-profiles-29421330-6vhh9\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.492945 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lww2l\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.498449 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.498739 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:42.998727492 +0000 UTC m=+148.239486011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.512700 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.573351 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq8x9\" (UniqueName: \"kubernetes.io/projected/348109cc-1761-4c83-b4fb-cdf2e5aa7e1a-kube-api-access-vq8x9\") pod \"ingress-canary-q9hbn\" (UID: \"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a\") " pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.573458 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lf44\" (UID: \"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.574106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12622b53-0ff2-40ac-9e7e-78b3dcaf8e38-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xn9f7\" (UID: \"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.576736 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2mhm\" (UniqueName: \"kubernetes.io/projected/f0fc13ea-940c-43f5-b256-245017e6664f-kube-api-access-v2mhm\") pod \"etcd-operator-b45778765-mwlr8\" (UID: \"f0fc13ea-940c-43f5-b256-245017e6664f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.587179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pknf9\" (UniqueName: \"kubernetes.io/projected/25b57e6c-de38-4c55-9960-b86dccd6b3e4-kube-api-access-pknf9\") pod \"machine-config-server-67vrh\" (UID: \"25b57e6c-de38-4c55-9960-b86dccd6b3e4\") " pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.599076 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.599260 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.099242379 +0000 UTC m=+148.340000888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.599376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.599674 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.09966294 +0000 UTC m=+148.340421459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.603054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.610047 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvf5s\" (UniqueName: \"kubernetes.io/projected/365d6032-3c70-47f9-ae54-632097bd0a47-kube-api-access-xvf5s\") pod \"machine-config-controller-84d6567774-gpl9q\" (UID: \"365d6032-3c70-47f9-ae54-632097bd0a47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.637845 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnljg\" (UniqueName: \"kubernetes.io/projected/e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84-kube-api-access-wnljg\") pod \"router-default-5444994796-ftbnh\" (UID: \"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84\") " pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.650154 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqhdp\" (UniqueName: \"kubernetes.io/projected/4d96f97a-35bd-4d9f-998b-0b1316c5e39d-kube-api-access-lqhdp\") pod \"migrator-59844c95c7-8hkwp\" (UID: \"4d96f97a-35bd-4d9f-998b-0b1316c5e39d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.650208 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.664352 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.666848 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.676186 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.682271 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.696801 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.703559 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.704033 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.204014127 +0000 UTC m=+148.444772646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.704138 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.705006 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.204998262 +0000 UTC m=+148.445756781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.706698 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.717979 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.756083 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-67vrh" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.770096 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.776799 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wrnx\" (UniqueName: \"kubernetes.io/projected/6d970f00-0a9d-4259-b21a-503e1e0bceb2-kube-api-access-9wrnx\") pod \"packageserver-d55dfcdfc-7rg7c\" (UID: \"6d970f00-0a9d-4259-b21a-503e1e0bceb2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.781656 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq6k9\" (UniqueName: \"kubernetes.io/projected/3dc28a15-7069-4b05-aced-372d68a397ff-kube-api-access-rq6k9\") pod \"catalog-operator-68c6474976-gr925\" (UID: \"3dc28a15-7069-4b05-aced-372d68a397ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.782284 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.788784 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cnbd\" (UniqueName: \"kubernetes.io/projected/a3d2442f-1a13-4e9a-903c-6ed4af79ca24-kube-api-access-5cnbd\") pod \"package-server-manager-789f6589d5-lqkzj\" (UID: \"a3d2442f-1a13-4e9a-903c-6ed4af79ca24\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.795595 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9gvr\" (UniqueName: \"kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr\") pod \"marketplace-operator-79b997595-sfxhc\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.805369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.806068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.806339 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6sgd\" (UniqueName: \"kubernetes.io/projected/a7df0970-4671-4495-8c49-b1c2c096885e-kube-api-access-g6sgd\") pod \"service-ca-9c57cc56f-9tllw\" (UID: \"a7df0970-4671-4495-8c49-b1c2c096885e\") " pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:42 crc kubenswrapper[4770]: E1209 11:33:42.807186 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.307163882 +0000 UTC m=+148.547922401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.807604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmkn5\" (UniqueName: \"kubernetes.io/projected/6684f72f-acbf-4c83-a6d6-1aa9e32a9941-kube-api-access-bmkn5\") pod \"multus-admission-controller-857f4d67dd-64jcp\" (UID: \"6684f72f-acbf-4c83-a6d6-1aa9e32a9941\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.812108 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl9x7\" (UniqueName: \"kubernetes.io/projected/bd1aa89d-775c-47be-a2ba-bf592371938d-kube-api-access-pl9x7\") pod \"dns-default-9xgcj\" (UID: \"bd1aa89d-775c-47be-a2ba-bf592371938d\") " pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.821486 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g242h\" (UniqueName: \"kubernetes.io/projected/3cd3eef7-d4d9-4f4c-aba7-d445e622383b-kube-api-access-g242h\") pod \"csi-hostpathplugin-vpb8g\" (UID: \"3cd3eef7-d4d9-4f4c-aba7-d445e622383b\") " pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.835177 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.860854 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:42 crc kubenswrapper[4770]: I1209 11:33:42.876183 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:42.976547 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:42.977501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:42.977981 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.477965843 +0000 UTC m=+148.718724362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:42.982019 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.023422 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.060700 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" event={"ID":"d0c869f7-6828-4624-be3c-1449ffc2e514","Type":"ContainerStarted","Data":"72f5eacd92b35054ac9647508cb20e6a6f4a87d36e228914aaa56fdf75a46fbc"} Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.078813 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.079057 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.079217 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.57919387 +0000 UTC m=+148.819952399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.079317 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.079606 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.57959448 +0000 UTC m=+148.820352999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.088031 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.193394 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.193706 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.693676296 +0000 UTC m=+148.934434815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.193833 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.194207 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.694196359 +0000 UTC m=+148.934954878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.256696 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.295135 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.295931 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.795892318 +0000 UTC m=+149.036650887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.301941 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q9hbn" Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.396348 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.397044 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.897031462 +0000 UTC m=+149.137789981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.498212 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.498604 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:43.998580926 +0000 UTC m=+149.239339445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.599651 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.600250 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.100222763 +0000 UTC m=+149.340981282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.724149 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.724633 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.224613003 +0000 UTC m=+149.465371522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.827266 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:43 crc kubenswrapper[4770]: E1209 11:33:43.827706 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.327691097 +0000 UTC m=+149.568449616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:43 crc kubenswrapper[4770]: I1209 11:33:43.931000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:43.931528 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.43150522 +0000 UTC m=+149.672263739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.034735 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.035426 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.535413435 +0000 UTC m=+149.776171954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.075558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" event={"ID":"d0c869f7-6828-4624-be3c-1449ffc2e514","Type":"ContainerStarted","Data":"eb949b7af9434b977069e8c6a15097bb9c0dae554d2e39b9ec99ffc8e16fa6c9"} Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.077190 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-67vrh" event={"ID":"25b57e6c-de38-4c55-9960-b86dccd6b3e4","Type":"ContainerStarted","Data":"18b448d1641fb5681396c853d018620534799eb352a1204bd1310195bb553220"} Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.089372 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ftbnh" event={"ID":"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84","Type":"ContainerStarted","Data":"8f38e818c3e6dd108147b41a4c8f3c739c14dfa9f0c6c0fe6435b4cd9fc16f02"} Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.136805 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.137186 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.637160104 +0000 UTC m=+149.877918623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.137593 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.137968 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.637960505 +0000 UTC m=+149.878719014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.238289 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.238699 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.738683109 +0000 UTC m=+149.979441628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.344019 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.344383 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:44.844369079 +0000 UTC m=+150.085127598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.521660 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.522062 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.022045485 +0000 UTC m=+150.262804004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.623400 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.623803 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.123788345 +0000 UTC m=+150.364546864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.768960 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.769678 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.269641826 +0000 UTC m=+150.510400355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:44 crc kubenswrapper[4770]: I1209 11:33:44.986184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:44 crc kubenswrapper[4770]: E1209 11:33:44.986583 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.48656831 +0000 UTC m=+150.727326829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.087732 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.088268 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.588246497 +0000 UTC m=+150.829005016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.094649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-67vrh" event={"ID":"25b57e6c-de38-4c55-9960-b86dccd6b3e4","Type":"ContainerStarted","Data":"0c58cc61a24e5729ad1e69194ef30bf5b41ffe19deda7e43e608a1da3d20b4f3"} Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.096806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ftbnh" event={"ID":"e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84","Type":"ContainerStarted","Data":"83f6fa383e98fb726fc15452b83404fb10947949204f357d2ba4bad6e48fb8d0"} Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.138337 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cv4wq" podStartSLOduration=127.138314022 podStartE2EDuration="2m7.138314022s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:44.521537262 +0000 UTC m=+149.762295781" watchObservedRunningTime="2025-12-09 11:33:45.138314022 +0000 UTC m=+150.379072541" Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.172488 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-67vrh" podStartSLOduration=6.172464918 podStartE2EDuration="6.172464918s" podCreationTimestamp="2025-12-09 11:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:45.141448752 +0000 UTC m=+150.382207271" watchObservedRunningTime="2025-12-09 11:33:45.172464918 +0000 UTC m=+150.413223437" Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.173095 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ftbnh" podStartSLOduration=127.173066813 podStartE2EDuration="2m7.173066813s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:45.169777079 +0000 UTC m=+150.410535608" watchObservedRunningTime="2025-12-09 11:33:45.173066813 +0000 UTC m=+150.413825342" Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.240959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.243294 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.743272134 +0000 UTC m=+150.984030863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.258648 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.341854 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.342637 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.842410936 +0000 UTC m=+151.083171495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.446652 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.447002 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:45.946990109 +0000 UTC m=+151.187748618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.547574 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.548199 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.048173124 +0000 UTC m=+151.288931643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.563218 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.563279 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.667392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.667777 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.16772164 +0000 UTC m=+151.408480159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.770564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.771044 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.27102598 +0000 UTC m=+151.511784499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.872249 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.872734 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.372712138 +0000 UTC m=+151.613470657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.911541 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f"] Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.916121 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z"] Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.973304 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.973956 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.473920803 +0000 UTC m=+151.714679332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:45 crc kubenswrapper[4770]: I1209 11:33:45.974184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:45 crc kubenswrapper[4770]: E1209 11:33:45.974756 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.474737134 +0000 UTC m=+151.715495653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.023486 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.030622 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.043252 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-42jsz"] Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.050053 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43156f0f_537c_4483_827e_d02941e9a274.slice/crio-de1fd3f3292070ac131a9b1ea47a29925af6d413eb344a9237e96590a7b70e45 WatchSource:0}: Error finding container de1fd3f3292070ac131a9b1ea47a29925af6d413eb344a9237e96590a7b70e45: Status 404 returned error can't find the container with id de1fd3f3292070ac131a9b1ea47a29925af6d413eb344a9237e96590a7b70e45 Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.050183 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.077882 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.078453 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.578423684 +0000 UTC m=+151.819182203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.095273 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.100040 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wdxh"] Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.102596 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae983b39_dbad_4db0_97f9_2bab5bb3ef36.slice/crio-7f94bda264ece259c5c7034cd196a5f7d2042c36a2470fdbd80f2cbd6335319d WatchSource:0}: Error finding container 7f94bda264ece259c5c7034cd196a5f7d2042c36a2470fdbd80f2cbd6335319d: Status 404 returned error can't find the container with id 7f94bda264ece259c5c7034cd196a5f7d2042c36a2470fdbd80f2cbd6335319d Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.106842 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode69f8571_c321_4ff0_9fd1_fb67b1176230.slice/crio-cbae5acb1cf16f8d54afefe42299cad08c4c0930264a2eec5a8c2d72114cc5cb WatchSource:0}: Error finding container cbae5acb1cf16f8d54afefe42299cad08c4c0930264a2eec5a8c2d72114cc5cb: Status 404 returned error can't find the container with id cbae5acb1cf16f8d54afefe42299cad08c4c0930264a2eec5a8c2d72114cc5cb Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.107131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" event={"ID":"25c0c5e8-78eb-427b-88f9-e02365a6afc1","Type":"ContainerStarted","Data":"b56d9e861c3faf920a9edb85cce1fe536c41b5790c9c41c6e142164e42cac599"} Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.114300 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" event={"ID":"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c","Type":"ContainerStarted","Data":"2fb509dc34d6b09884ea4fe95caf0da23ad5eba2a5078b6f2e41e7604fb467b4"} Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.115419 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44f6s"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.116359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" event={"ID":"43156f0f-537c-4483-827e-d02941e9a274","Type":"ContainerStarted","Data":"de1fd3f3292070ac131a9b1ea47a29925af6d413eb344a9237e96590a7b70e45"} Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.117363 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" event={"ID":"bd6aba04-bce5-4231-b2c5-c3b574409d90","Type":"ContainerStarted","Data":"8295b130c99283bb322cfb99fcc1fd36370d1449e1064592202c9b9ef35571c7"} Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.118616 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" event={"ID":"c86e7caf-f2ca-4006-a807-e055f0597bf9","Type":"ContainerStarted","Data":"97bfc79d5119206efc585701ae542e0aff862ed74d271dd655cd15e072b35e00"} Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.118700 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xwjv4"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.122108 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-j4njp"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.123319 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r"] Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.168344 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc891628b_1ab0_498d_8792_60029bbff6b3.slice/crio-cb6c68f4f58f65cf13f9d5eeed91f354aa6741a443f89d3557bdb20775722de4 WatchSource:0}: Error finding container cb6c68f4f58f65cf13f9d5eeed91f354aa6741a443f89d3557bdb20775722de4: Status 404 returned error can't find the container with id cb6c68f4f58f65cf13f9d5eeed91f354aa6741a443f89d3557bdb20775722de4 Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.179320 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.181598 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.681301562 +0000 UTC m=+151.922060141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.264769 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:46 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:46 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:46 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.264818 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.280309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.280575 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.780524457 +0000 UTC m=+152.021283016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.306146 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5nk5g"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.330721 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9tllw"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.331674 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.336861 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9xgcj"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.339557 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.355137 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.360392 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.383389 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.383734 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.883721184 +0000 UTC m=+152.124479703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.384240 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.402299 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.425435 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.483986 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.484134 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.984113509 +0000 UTC m=+152.224872038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.487011 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.487756 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:46.987739112 +0000 UTC m=+152.228497631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.564858 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vpb8g"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.588061 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.588326 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.088295681 +0000 UTC m=+152.329054200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.604959 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.606254 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-64jcp"] Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.645008 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda58a8c2f_222c_41c7_8189_2610471f69df.slice/crio-e36691de5cec3ec529f9330f3f666c4a6053832ac469d68e8ced081b38cb9a0f WatchSource:0}: Error finding container e36691de5cec3ec529f9330f3f666c4a6053832ac469d68e8ced081b38cb9a0f: Status 404 returned error can't find the container with id e36691de5cec3ec529f9330f3f666c4a6053832ac469d68e8ced081b38cb9a0f Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.651706 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.689582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.690069 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.190048831 +0000 UTC m=+152.430807350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: W1209 11:33:46.730494 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6684f72f_acbf_4c83_a6d6_1aa9e32a9941.slice/crio-a725dae661aec70d917a5f3ccb7ca92bfa6993d95a39a24ede6fe62fee639109 WatchSource:0}: Error finding container a725dae661aec70d917a5f3ccb7ca92bfa6993d95a39a24ede6fe62fee639109: Status 404 returned error can't find the container with id a725dae661aec70d917a5f3ccb7ca92bfa6993d95a39a24ede6fe62fee639109 Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.792387 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.792584 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.29255598 +0000 UTC m=+152.533314499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.792721 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.793323 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.293314319 +0000 UTC m=+152.534072838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.896509 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:46 crc kubenswrapper[4770]: E1209 11:33:46.897483 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.39745518 +0000 UTC m=+152.638213699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.934658 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.946769 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8t95q"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.974315 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c"] Dec 09 11:33:46 crc kubenswrapper[4770]: I1209 11:33:46.992169 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.000289 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.000968 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.500951825 +0000 UTC m=+152.741710344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.018925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwlr8"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.028963 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd"] Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.048235 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32fdb059_5acb_446d_98ac_8239919a90b7.slice/crio-01af3ff6baba12e91fc31481ab23108d38243b9ab9ed19aee8dcd84065337ccb WatchSource:0}: Error finding container 01af3ff6baba12e91fc31481ab23108d38243b9ab9ed19aee8dcd84065337ccb: Status 404 returned error can't find the container with id 01af3ff6baba12e91fc31481ab23108d38243b9ab9ed19aee8dcd84065337ccb Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.049882 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d970f00_0a9d_4259_b21a_503e1e0bceb2.slice/crio-72eca52e5b6836057f9f4345556dfcb78aa82a806aff97d7c2ae6fc0f3f99052 WatchSource:0}: Error finding container 72eca52e5b6836057f9f4345556dfcb78aa82a806aff97d7c2ae6fc0f3f99052: Status 404 returned error can't find the container with id 72eca52e5b6836057f9f4345556dfcb78aa82a806aff97d7c2ae6fc0f3f99052 Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.051815 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94d68ac6_09be_4ca6_8221_f508337ec0c7.slice/crio-66972805ad510c9cf39be2451664b16c1c831da2b50c9a71e2968d642ca119b5 WatchSource:0}: Error finding container 66972805ad510c9cf39be2451664b16c1c831da2b50c9a71e2968d642ca119b5: Status 404 returned error can't find the container with id 66972805ad510c9cf39be2451664b16c1c831da2b50c9a71e2968d642ca119b5 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.056308 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb"] Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.071477 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0fc13ea_940c_43f5_b256_245017e6664f.slice/crio-a0cb8a6d46cac6e78b99ac850dd4a690628db09a51efabaa9efda2bdf562e095 WatchSource:0}: Error finding container a0cb8a6d46cac6e78b99ac850dd4a690628db09a51efabaa9efda2bdf562e095: Status 404 returned error can't find the container with id a0cb8a6d46cac6e78b99ac850dd4a690628db09a51efabaa9efda2bdf562e095 Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.079327 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e08edb8_6c28_49a2_ba80_7e3703f48f5f.slice/crio-97dd1f127cf0728157a00bf4079144992bb4bbcdc1e7327f942b3a2568ca289d WatchSource:0}: Error finding container 97dd1f127cf0728157a00bf4079144992bb4bbcdc1e7327f942b3a2568ca289d: Status 404 returned error can't find the container with id 97dd1f127cf0728157a00bf4079144992bb4bbcdc1e7327f942b3a2568ca289d Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.087398 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.101789 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.102343 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.602267934 +0000 UTC m=+152.843026453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.112377 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fsglb"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.143870 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q9hbn"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.148029 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.153635 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgdc" event={"ID":"9357bac1-cc23-4f63-8d12-458305a47f77","Type":"ContainerStarted","Data":"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.153687 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgdc" event={"ID":"9357bac1-cc23-4f63-8d12-458305a47f77","Type":"ContainerStarted","Data":"de2c561374c6f596f237b4f07579f0d4b28314d6b3fecf04ea3ea817dcc50804"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.157031 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9xgcj" event={"ID":"bd1aa89d-775c-47be-a2ba-bf592371938d","Type":"ContainerStarted","Data":"560a0d7aa83be01528c8ca6ab7ce6c5313d0184ab1ec76d6555e5111d7b834dd"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.165145 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" event={"ID":"7ed604a1-90f7-4f46-a4a3-8e097b02c922","Type":"ContainerStarted","Data":"71be2c13ea2570302d8866dc620b7948c7b52fbedc2f7d42e95036ec25e9f051"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.165198 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" event={"ID":"7ed604a1-90f7-4f46-a4a3-8e097b02c922","Type":"ContainerStarted","Data":"6e594a6c3e80f78329c20d412832ee7e019678cc1e95ec343d022538e10fd46b"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.174669 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" event={"ID":"c0d5b7a7-6e5f-430e-863a-a18ac009ee3c","Type":"ContainerStarted","Data":"8ff69ba0e5471e5effca3c01fdccc6ba303e4d16d89ff3301c25fd70145effa6"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.184182 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" event={"ID":"4d96f97a-35bd-4d9f-998b-0b1316c5e39d","Type":"ContainerStarted","Data":"9bd98be9494228d51fa4928c544b4c53e70f258b8b5df86c6c63105fb8b716ae"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.184228 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" event={"ID":"4d96f97a-35bd-4d9f-998b-0b1316c5e39d","Type":"ContainerStarted","Data":"df6765161240241b655e28525f11ff15177d70287e1fb2cddc1b4f1ef5e3a194"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.188266 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mhgdc" podStartSLOduration=129.188246969 podStartE2EDuration="2m9.188246969s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.1832252 +0000 UTC m=+152.423983719" watchObservedRunningTime="2025-12-09 11:33:47.188246969 +0000 UTC m=+152.429005488" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.188977 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7"] Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.189410 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" event={"ID":"f0fc13ea-940c-43f5-b256-245017e6664f","Type":"ContainerStarted","Data":"a0cb8a6d46cac6e78b99ac850dd4a690628db09a51efabaa9efda2bdf562e095"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.199256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" event={"ID":"ff350dff-877d-4fac-8ed7-322c0e8894e0","Type":"ContainerStarted","Data":"380918ecfd2f605c8a9a3f4cf11838b014e6d05d715ff43cffc62a3b62fa037a"} Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.204214 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod412ed7a7_4ea3_4daf_8978_a70cf1702c1e.slice/crio-3cad12d8a4babc47323975c37a54ca1a7c6bfda6a07fede04547ef4f733bbf3f WatchSource:0}: Error finding container 3cad12d8a4babc47323975c37a54ca1a7c6bfda6a07fede04547ef4f733bbf3f: Status 404 returned error can't find the container with id 3cad12d8a4babc47323975c37a54ca1a7c6bfda6a07fede04547ef4f733bbf3f Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.205111 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.205450 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.70543881 +0000 UTC m=+152.946197329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.207606 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" event={"ID":"e1a5a731-6de5-4cfa-abfd-b68487c0b855","Type":"ContainerStarted","Data":"7b15bc9df0119f263b45a27ce813b40eb30d2cc433a08532b16a9372aa80fff2"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.207647 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" event={"ID":"e1a5a731-6de5-4cfa-abfd-b68487c0b855","Type":"ContainerStarted","Data":"2d434e29bbb092e2cff42ef58636ae91041b9528a79dccc3044474b868811b92"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.208491 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.217349 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b6p2f" podStartSLOduration=129.217332045 podStartE2EDuration="2m9.217332045s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.21517648 +0000 UTC m=+152.455935009" watchObservedRunningTime="2025-12-09 11:33:47.217332045 +0000 UTC m=+152.458090564" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.230129 4770 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2wdxh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.230205 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.230274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" event={"ID":"4e08edb8-6c28-49a2-ba80-7e3703f48f5f","Type":"ContainerStarted","Data":"97dd1f127cf0728157a00bf4079144992bb4bbcdc1e7327f942b3a2568ca289d"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.242072 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ea4969913f8f976e57cf27258c4ce1db9d21169e761987564a8d3c28fbc91bcf"} Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.254919 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode26ca10a_d6e1_4b64_a7b9_5b4daceae7a9.slice/crio-c0a8e26f42dfa7b00a6db1502b7df2cc83c015d908e701b84bb445bb1eca2e52 WatchSource:0}: Error finding container c0a8e26f42dfa7b00a6db1502b7df2cc83c015d908e701b84bb445bb1eca2e52: Status 404 returned error can't find the container with id c0a8e26f42dfa7b00a6db1502b7df2cc83c015d908e701b84bb445bb1eca2e52 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.255179 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerStarted","Data":"1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.255267 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerStarted","Data":"cbae5acb1cf16f8d54afefe42299cad08c4c0930264a2eec5a8c2d72114cc5cb"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.256677 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.265804 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" podStartSLOduration=129.265786798 podStartE2EDuration="2m9.265786798s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.26544955 +0000 UTC m=+152.506208059" watchObservedRunningTime="2025-12-09 11:33:47.265786798 +0000 UTC m=+152.506545317" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.267246 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.267289 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.272811 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:47 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:47 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:47 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.272856 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.273809 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" event={"ID":"94d68ac6-09be-4ca6-8221-f508337ec0c7","Type":"ContainerStarted","Data":"66972805ad510c9cf39be2451664b16c1c831da2b50c9a71e2968d642ca119b5"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.275657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" event={"ID":"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e","Type":"ContainerStarted","Data":"95c68d76fc98344ccdf0ba7faf5e98ddba9b6e83f50b2553e19dd2e6f703ea40"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.284607 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" event={"ID":"3cd3eef7-d4d9-4f4c-aba7-d445e622383b","Type":"ContainerStarted","Data":"1f01d6132a7f25d0b95d26bc8d5a5b0d73455dff5f11bbf26162bf0ec2dc1f9d"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.291916 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xwjv4" podStartSLOduration=129.291875527 podStartE2EDuration="2m9.291875527s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.290889612 +0000 UTC m=+152.531648131" watchObservedRunningTime="2025-12-09 11:33:47.291875527 +0000 UTC m=+152.532634046" Dec 09 11:33:47 crc kubenswrapper[4770]: W1209 11:33:47.292914 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod348109cc_1761_4c83_b4fb_cdf2e5aa7e1a.slice/crio-0263cea965d18ef9728894e0549986606fb6444e44e037a46c419fa674583eb5 WatchSource:0}: Error finding container 0263cea965d18ef9728894e0549986606fb6444e44e037a46c419fa674583eb5: Status 404 returned error can't find the container with id 0263cea965d18ef9728894e0549986606fb6444e44e037a46c419fa674583eb5 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.295532 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-44f6s" event={"ID":"ae983b39-dbad-4db0-97f9-2bab5bb3ef36","Type":"ContainerStarted","Data":"37541d23392c77be29f1c3fda5366f6e1f34c9e9aebafdd96c641df9347f79a5"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.295588 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-44f6s" event={"ID":"ae983b39-dbad-4db0-97f9-2bab5bb3ef36","Type":"ContainerStarted","Data":"7f94bda264ece259c5c7034cd196a5f7d2042c36a2470fdbd80f2cbd6335319d"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.296085 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.383332 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.383864 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.883827906 +0000 UTC m=+153.124586425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.389128 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.392979 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:47.89295446 +0000 UTC m=+153.133712979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.397875 4770 generic.go:334] "Generic (PLEG): container finished" podID="e215f2a5-db5f-4828-8824-bd30bc0f6c4d" containerID="ca09b336c15d5e0b5355edea0cb796415707b27d9975e77c6f513843a74bbd30" exitCode=0 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.413555 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" event={"ID":"e215f2a5-db5f-4828-8824-bd30bc0f6c4d","Type":"ContainerDied","Data":"ca09b336c15d5e0b5355edea0cb796415707b27d9975e77c6f513843a74bbd30"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.413598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" event={"ID":"e215f2a5-db5f-4828-8824-bd30bc0f6c4d","Type":"ContainerStarted","Data":"14fbaa370e9b00a41fe80d34a0458d84dfd31a9ce82c0db7d425481227511295"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.548762 4770 generic.go:334] "Generic (PLEG): container finished" podID="25c0c5e8-78eb-427b-88f9-e02365a6afc1" containerID="c6b2a5acb989a42f114422a4efddeb51daffcd4fb5ae392418039a2651410e9f" exitCode=0 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.548863 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" event={"ID":"25c0c5e8-78eb-427b-88f9-e02365a6afc1","Type":"ContainerDied","Data":"c6b2a5acb989a42f114422a4efddeb51daffcd4fb5ae392418039a2651410e9f"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.550830 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.552178 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.052154483 +0000 UTC m=+153.292913052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.576376 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ea91fad41f788c9316633d6098c331f8a2a11de4f95dcaffd94057447c68838f"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.578517 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" event={"ID":"a3d2442f-1a13-4e9a-903c-6ed4af79ca24","Type":"ContainerStarted","Data":"13fea1bcb5c834cfcff8489d94388c15517441d65be4164e010e1532b9a3c4bf"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.579485 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"421831589f512eda05ae6e8f4aa76e8fdee4222d64c4f0d0f3d191937c0cd698"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.580469 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" event={"ID":"6684f72f-acbf-4c83-a6d6-1aa9e32a9941","Type":"ContainerStarted","Data":"a725dae661aec70d917a5f3ccb7ca92bfa6993d95a39a24ede6fe62fee639109"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.582211 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" event={"ID":"bd6aba04-bce5-4231-b2c5-c3b574409d90","Type":"ContainerStarted","Data":"4bfd8733cc57675215c929e31b7cd61e4c7d8187f617d8dc5ea1142f71a0b475"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.587751 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.589543 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-44f6s" podStartSLOduration=129.589520661 podStartE2EDuration="2m9.589520661s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.418662739 +0000 UTC m=+152.659421258" watchObservedRunningTime="2025-12-09 11:33:47.589520661 +0000 UTC m=+152.830279190" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.591996 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" event={"ID":"43156f0f-537c-4483-827e-d02941e9a274","Type":"ContainerStarted","Data":"cffc514a0ccf2a7d346c21a6fe441e5844726482a6b296c61a5d9c9ea5b87baa"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.622249 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.633357 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" event={"ID":"45e8e8a9-2925-41c1-b8f5-7aafde0d700d","Type":"ContainerStarted","Data":"3d9812e763ec7bd8d65fdef33118ac00781bc0c3419015d12b1ff5102c1c3181"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.639058 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" event={"ID":"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77","Type":"ContainerStarted","Data":"c7312ee147831b38fbef1a2143cf6b2ebacd7565f9666615b7cab49f236bbe28"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.643042 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" event={"ID":"1f021879-eaf1-4d8b-8d8f-bd91525114c3","Type":"ContainerStarted","Data":"3326430c23981f784398a98c4fb14974fd8aea7a7a84d382883b48e02c47c406"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.648949 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" event={"ID":"32fdb059-5acb-446d-98ac-8239919a90b7","Type":"ContainerStarted","Data":"01af3ff6baba12e91fc31481ab23108d38243b9ab9ed19aee8dcd84065337ccb"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.650467 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" event={"ID":"a7df0970-4671-4495-8c49-b1c2c096885e","Type":"ContainerStarted","Data":"7207b27f35805da21b0c089845d4ecfdc1acc4b9c7c4592cea063e5b08ebffa2"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.652445 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.657344 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.157330281 +0000 UTC m=+153.398088810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.670305 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" event={"ID":"c86e7caf-f2ca-4006-a807-e055f0597bf9","Type":"ContainerStarted","Data":"bcc42931cb9a3303ba84cbbec94a6141bf6a61068dadc771d8c55e81961d2229"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.731283 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" podStartSLOduration=129.731262787 podStartE2EDuration="2m9.731262787s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.671402122 +0000 UTC m=+152.912160641" watchObservedRunningTime="2025-12-09 11:33:47.731262787 +0000 UTC m=+152.972021306" Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.741037 4770 generic.go:334] "Generic (PLEG): container finished" podID="c891628b-1ab0-498d-8792-60029bbff6b3" containerID="a45e12f4820824418d13cf083ad906c1a952efe007b96969ec681f9f772560e6" exitCode=0 Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.741122 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" event={"ID":"c891628b-1ab0-498d-8792-60029bbff6b3","Type":"ContainerDied","Data":"a45e12f4820824418d13cf083ad906c1a952efe007b96969ec681f9f772560e6"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.741159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" event={"ID":"c891628b-1ab0-498d-8792-60029bbff6b3","Type":"ContainerStarted","Data":"cb6c68f4f58f65cf13f9d5eeed91f354aa6741a443f89d3557bdb20775722de4"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.753505 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.753846 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.253827196 +0000 UTC m=+153.494585715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.760277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" event={"ID":"a58a8c2f-222c-41c7-8189-2610471f69df","Type":"ContainerStarted","Data":"e36691de5cec3ec529f9330f3f666c4a6053832ac469d68e8ced081b38cb9a0f"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.761729 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" event={"ID":"6d970f00-0a9d-4259-b21a-503e1e0bceb2","Type":"ContainerStarted","Data":"72eca52e5b6836057f9f4345556dfcb78aa82a806aff97d7c2ae6fc0f3f99052"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.762705 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" event={"ID":"3dc28a15-7069-4b05-aced-372d68a397ff","Type":"ContainerStarted","Data":"3e2854f22a5fdafeeaec4dbbd67bbe75f79c6a0726605e0543c824229e8c03ca"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.771119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" event={"ID":"76a56c07-a692-4268-b1d1-e7535e738ad3","Type":"ContainerStarted","Data":"1d2bfd5072fefca13246038090585097ce723bf40fa4b47b91abcaabfd6ce260"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.780922 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" event={"ID":"8e1f7c2c-c165-409d-9e75-a22656ae8c12","Type":"ContainerStarted","Data":"a3fda7b671ffc158617512f0aafd7583e83f80383900bcd5bc6ae2460b6d3924"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.781025 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" event={"ID":"8e1f7c2c-c165-409d-9e75-a22656ae8c12","Type":"ContainerStarted","Data":"3bbcaa8ea571d69fcd57a2d300db6fd7c37363c2365eedb426aa5c5da63a18b6"} Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.858128 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.858730 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.358690925 +0000 UTC m=+153.599449444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:47 crc kubenswrapper[4770]: I1209 11:33:47.966002 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:47 crc kubenswrapper[4770]: E1209 11:33:47.968349 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.468314437 +0000 UTC m=+153.709072956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.027470 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" podStartSLOduration=130.027449564 podStartE2EDuration="2m10.027449564s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:48.025400132 +0000 UTC m=+153.266158651" watchObservedRunningTime="2025-12-09 11:33:48.027449564 +0000 UTC m=+153.268208093" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.027747 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlkvz" podStartSLOduration=130.027740552 podStartE2EDuration="2m10.027740552s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:47.949091624 +0000 UTC m=+153.189850143" watchObservedRunningTime="2025-12-09 11:33:48.027740552 +0000 UTC m=+153.268499071" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.068609 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.069241 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.569221064 +0000 UTC m=+153.809979593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.166515 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qr7cm" podStartSLOduration=130.166482509 podStartE2EDuration="2m10.166482509s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:48.112648188 +0000 UTC m=+153.353406707" watchObservedRunningTime="2025-12-09 11:33:48.166482509 +0000 UTC m=+153.407241028" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.189402 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.189711 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.689673344 +0000 UTC m=+153.930431863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.189766 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.190574 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.690551986 +0000 UTC m=+153.931310505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.276251 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" podStartSLOduration=129.276230863 podStartE2EDuration="2m9.276230863s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:48.188730949 +0000 UTC m=+153.429489468" watchObservedRunningTime="2025-12-09 11:33:48.276230863 +0000 UTC m=+153.516989382" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.278096 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-42jsz" podStartSLOduration=130.278088781 podStartE2EDuration="2m10.278088781s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:48.275521475 +0000 UTC m=+153.516279994" watchObservedRunningTime="2025-12-09 11:33:48.278088781 +0000 UTC m=+153.518847300" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.305659 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.306085 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.806062299 +0000 UTC m=+154.046820818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.393836 4770 patch_prober.go:28] interesting pod/console-operator-58897d9998-44f6s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.394196 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-44f6s" podUID="ae983b39-dbad-4db0-97f9-2bab5bb3ef36" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.400369 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:48 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:48 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:48 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.400433 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.412286 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.412709 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:48.912693694 +0000 UTC m=+154.153452213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.525009 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.525396 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.025369854 +0000 UTC m=+154.266128373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.635084 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.635392 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.135379185 +0000 UTC m=+154.376137704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.736677 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.737209 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.237180826 +0000 UTC m=+154.477939345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.799657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" event={"ID":"ff350dff-877d-4fac-8ed7-322c0e8894e0","Type":"ContainerStarted","Data":"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.801037 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.803032 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5e1f0b4371704bbcae9fdf00154d2b89bdd032d25801e11a545344c5a9524bce"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.806465 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-89tfr" event={"ID":"6c5d9c5f-9b56-4e7d-b7e6-5132c0255c77","Type":"ContainerStarted","Data":"885b39b3f024463445f9152c85cf306b46501e9c22a5c7ff50028fe1196485bb"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.808858 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" event={"ID":"1f021879-eaf1-4d8b-8d8f-bd91525114c3","Type":"ContainerStarted","Data":"e73740f1468b7f75cc6c2726785d0fa79767d3b3302b9a6ea8fff4a1e57948e1"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.813377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"de560fc4ea2533b3b051e3fbb191c8042d1a36b1362302169962b15e0b71e1b5"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.838765 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:48 crc kubenswrapper[4770]: E1209 11:33:48.840754 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.340739972 +0000 UTC m=+154.581498491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.842425 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" event={"ID":"6684f72f-acbf-4c83-a6d6-1aa9e32a9941","Type":"ContainerStarted","Data":"47fb43db75b45880ece2e8f08fee641bfaa38d126d4f76d485f506c7a136e8af"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.844739 4770 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d8ftw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.844786 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.845315 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" event={"ID":"8a5cb1c7-5a1e-4cf9-a07b-381dda9b165e","Type":"ContainerStarted","Data":"c1cbe51d4eb45f61e81152ec8ec1637f0329944a278ff3edc27ff2785f094a12"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.894132 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" event={"ID":"4d96f97a-35bd-4d9f-998b-0b1316c5e39d","Type":"ContainerStarted","Data":"9625579ad2afe84375f641768d43e5d582d0093975a142af3953f4c5762cbf4d"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.913513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" event={"ID":"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f","Type":"ContainerStarted","Data":"05871e2814e0a333a4be273f3b40d37d099e667f3b3661ad2015889c5adf7dc4"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.923098 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"263e9c4107aa12619e15bc6e040e0b859a10d05d31d3f11f8d5ba43fc9b22a2a"} Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.925217 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.938605 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ppj6k" podStartSLOduration=130.938587322 podStartE2EDuration="2m10.938587322s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:48.936457558 +0000 UTC m=+154.177216107" watchObservedRunningTime="2025-12-09 11:33:48.938587322 +0000 UTC m=+154.179345841" Dec 09 11:33:48 crc kubenswrapper[4770]: I1209 11:33:48.939415 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.021914 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.521874558 +0000 UTC m=+154.762633077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.024335 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podStartSLOduration=130.02430661 podStartE2EDuration="2m10.02430661s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.020448571 +0000 UTC m=+154.261207100" watchObservedRunningTime="2025-12-09 11:33:49.02430661 +0000 UTC m=+154.265065139" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.035675 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" event={"ID":"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9","Type":"ContainerStarted","Data":"c0a8e26f42dfa7b00a6db1502b7df2cc83c015d908e701b84bb445bb1eca2e52"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.042100 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.046312 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.546294714 +0000 UTC m=+154.787053233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.187322 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.188472 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.688212244 +0000 UTC m=+154.928970763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.207674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" event={"ID":"412ed7a7-4ea3-4daf-8978-a70cf1702c1e","Type":"ContainerStarted","Data":"3cad12d8a4babc47323975c37a54ca1a7c6bfda6a07fede04547ef4f733bbf3f"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.241678 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lf44" podStartSLOduration=131.241644625 podStartE2EDuration="2m11.241644625s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.241634905 +0000 UTC m=+154.482393434" watchObservedRunningTime="2025-12-09 11:33:49.241644625 +0000 UTC m=+154.482403144" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.316367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9tllw" event={"ID":"a7df0970-4671-4495-8c49-b1c2c096885e","Type":"ContainerStarted","Data":"aabb5dc81af3fc7499ea27337762843c0e624bf06f0bb42b1408c5dc8e7443d9"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.317676 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.330647 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" event={"ID":"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38","Type":"ContainerStarted","Data":"7afd747797e7f4e6a172100bb8be674c04240f6b8946d01af6c01b3c5a7bbc63"} Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.335592 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.818888186 +0000 UTC m=+155.059646705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.348283 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:49 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:49 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:49 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.348607 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.409870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" event={"ID":"45e8e8a9-2925-41c1-b8f5-7aafde0d700d","Type":"ContainerStarted","Data":"af57e8ec0c1efc2cec6b1ef6845a2b461e0a31e71fdaa4bfdfaed56c553f245e"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.410202 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" event={"ID":"76a56c07-a692-4268-b1d1-e7535e738ad3","Type":"ContainerStarted","Data":"c23d868c0819432b8ff4415cf541f28a7682f21fd5879a78ccffedfb29cd5f6a"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.418488 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.418722 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.918689656 +0000 UTC m=+155.159448175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.419222 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.419913 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:49.919868856 +0000 UTC m=+155.160627435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.421482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" event={"ID":"3dc28a15-7069-4b05-aced-372d68a397ff","Type":"ContainerStarted","Data":"6b556583af74846aca3e737139ac48de89bade18d8c8eaa3f530311c4504713c"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.422656 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.461035 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gr925 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.461095 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podUID="3dc28a15-7069-4b05-aced-372d68a397ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.520037 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podStartSLOduration=130.520015825 podStartE2EDuration="2m10.520015825s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.510777738 +0000 UTC m=+154.751536257" watchObservedRunningTime="2025-12-09 11:33:49.520015825 +0000 UTC m=+154.760774344" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.522049 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8hkwp" podStartSLOduration=131.522035946 podStartE2EDuration="2m11.522035946s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.487642144 +0000 UTC m=+154.728400663" watchObservedRunningTime="2025-12-09 11:33:49.522035946 +0000 UTC m=+154.762794465" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.531344 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.531716 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.031687714 +0000 UTC m=+155.272446233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.538534 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.540842 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.040824268 +0000 UTC m=+155.281582787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.691137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.691463 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.191447061 +0000 UTC m=+155.432205580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.693629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" event={"ID":"a58a8c2f-222c-41c7-8189-2610471f69df","Type":"ContainerStarted","Data":"9a8b0fd58c62095b9a4655efc588bb5b2956d0d3b3181e79a4cbc75831fa3c45"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.694250 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.739651 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cz9jk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.740425 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" podUID="a58a8c2f-222c-41c7-8189-2610471f69df" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.742355 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q9hbn" event={"ID":"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a","Type":"ContainerStarted","Data":"0263cea965d18ef9728894e0549986606fb6444e44e037a46c419fa674583eb5"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.778562 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9xgcj" event={"ID":"bd1aa89d-775c-47be-a2ba-bf592371938d","Type":"ContainerStarted","Data":"0a5fdc11925d95342dddb1541bf90fc3cef4cbff45159b9135dceace78b4d653"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.792237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.792792 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.292776221 +0000 UTC m=+155.533534750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.796927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" event={"ID":"7ed604a1-90f7-4f46-a4a3-8e097b02c922","Type":"ContainerStarted","Data":"16c5ba3dc42c83f8c01cfc4d77224ad9d1c3ec5b0a45039c8de5897bdb31416e"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.821642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" event={"ID":"365d6032-3c70-47f9-ae54-632097bd0a47","Type":"ContainerStarted","Data":"8c9defe3ddeba2bea0db461237c8443e633e738959428c8bca78225c2d3834e5"} Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.827204 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.827259 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.893175 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:49 crc kubenswrapper[4770]: E1209 11:33:49.894776 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.394756986 +0000 UTC m=+155.635515515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.896915 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" podStartSLOduration=130.896879221 podStartE2EDuration="2m10.896879221s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.892047757 +0000 UTC m=+155.132806276" watchObservedRunningTime="2025-12-09 11:33:49.896879221 +0000 UTC m=+155.137637730" Dec 09 11:33:49 crc kubenswrapper[4770]: I1209 11:33:49.966522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-44f6s" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.060296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.063335 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.563316909 +0000 UTC m=+155.804075618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.104351 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.104329 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5nk5g" podStartSLOduration=132.104310651 podStartE2EDuration="2m12.104310651s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:49.927721552 +0000 UTC m=+155.168480071" watchObservedRunningTime="2025-12-09 11:33:50.104310651 +0000 UTC m=+155.345069170" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.247991 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.248574 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.74853762 +0000 UTC m=+155.989296129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.272131 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:50 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:50 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:50 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.272538 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.387226 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.387658 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:50.887644098 +0000 UTC m=+156.128402617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.499714 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.500108 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.000078872 +0000 UTC m=+156.240837401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.503536 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.504084 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.004069874 +0000 UTC m=+156.244828393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.622119 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.622663 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.122621975 +0000 UTC m=+156.363380494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.770727 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.771066 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.271051131 +0000 UTC m=+156.511809650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.876843 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.877096 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.377080591 +0000 UTC m=+156.617839110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.922065 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" event={"ID":"412ed7a7-4ea3-4daf-8978-a70cf1702c1e","Type":"ContainerStarted","Data":"dbeb1e8e3d7b8665b89435c882a464bd455902ce5956f27d22d6eda616fdd502"} Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.924203 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" event={"ID":"6d970f00-0a9d-4259-b21a-503e1e0bceb2","Type":"ContainerStarted","Data":"57692ffae4f0ec09e2eaf18cecd2dcc7c9ba2651025b1fa1096b78eaff002f35"} Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.925373 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.927122 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7rg7c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.927163 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podUID="6d970f00-0a9d-4259-b21a-503e1e0bceb2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.929218 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" event={"ID":"365d6032-3c70-47f9-ae54-632097bd0a47","Type":"ContainerStarted","Data":"d2a83dbb9b4a14514d9359b05d61cfc4276cd19b5973d4ec165962cc8059b834"} Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931647 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931677 4770 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d8ftw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931704 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gr925 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931711 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931647 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cz9jk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931720 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931768 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" podUID="a58a8c2f-222c-41c7-8189-2610471f69df" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.931739 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podUID="3dc28a15-7069-4b05-aced-372d68a397ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 11:33:50 crc kubenswrapper[4770]: I1209 11:33:50.991477 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:50 crc kubenswrapper[4770]: E1209 11:33:50.992325 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.492294616 +0000 UTC m=+156.733053135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.104976 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.106614 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.606587848 +0000 UTC m=+156.847346367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.115256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.120398 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.620377021 +0000 UTC m=+156.861135550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.124875 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.124975 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.125106 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.125202 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.223856 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.224222 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.724203514 +0000 UTC m=+156.964962033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.237281 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.238159 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.255824 4770 patch_prober.go:28] interesting pod/console-f9d7485db-mhgdc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.255880 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mhgdc" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.258071 4770 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d8ftw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.258100 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.266142 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:51 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:51 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:51 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.266219 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.325377 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.325809 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.82579281 +0000 UTC m=+157.066551329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.426068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.426886 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:51.926857912 +0000 UTC m=+157.167616481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.527815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.528204 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.028189411 +0000 UTC m=+157.268947940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.599978 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.688573 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.689878 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.189857846 +0000 UTC m=+157.430616375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:51 crc kubenswrapper[4770]: I1209 11:33:51.866768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:51 crc kubenswrapper[4770]: E1209 11:33:51.867474 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.367461822 +0000 UTC m=+157.608220341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:51.999143 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podStartSLOduration=132.999123489 podStartE2EDuration="2m12.999123489s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:51.038741688 +0000 UTC m=+156.279500207" watchObservedRunningTime="2025-12-09 11:33:51.999123489 +0000 UTC m=+157.239882008" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.001776 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.007433 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.507409631 +0000 UTC m=+157.748168150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.109481 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.110407 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.610375662 +0000 UTC m=+157.851134181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.221039 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.222612 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.722558819 +0000 UTC m=+157.963317498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.262750 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:52 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:52 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:52 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.262827 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.294943 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" event={"ID":"76a56c07-a692-4268-b1d1-e7535e738ad3","Type":"ContainerStarted","Data":"378bce2e98e21e852d30761dee32441c477eca9d2c4c6e2e28348d8d318050c2"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.334550 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" event={"ID":"a3d2442f-1a13-4e9a-903c-6ed4af79ca24","Type":"ContainerStarted","Data":"0e6287b4c3d89d5b437b774a8633e7f07ccb0a2c4f3194abfe6539ca2fe191d2"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.336096 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.336514 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.836494492 +0000 UTC m=+158.077253021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.344839 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5vbfw" podStartSLOduration=134.344808175 podStartE2EDuration="2m14.344808175s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.344319392 +0000 UTC m=+157.585077931" watchObservedRunningTime="2025-12-09 11:33:52.344808175 +0000 UTC m=+157.585566694" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.364379 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" event={"ID":"4e08edb8-6c28-49a2-ba80-7e3703f48f5f","Type":"ContainerStarted","Data":"ed04eae26527309af718494f4690639ddf29fe4e0c609b2443cab5acc4f32a52"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.395035 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9xgcj" event={"ID":"bd1aa89d-775c-47be-a2ba-bf592371938d","Type":"ContainerStarted","Data":"9e0e77bf515255905c1a1f7a449cf081bd6715f37e151be2c33114f322c28efd"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.396117 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9xgcj" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.418310 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" event={"ID":"3cd3eef7-d4d9-4f4c-aba7-d445e622383b","Type":"ContainerStarted","Data":"9f5b314242b541a7868b81a30a40f0803a7a1ccc7f52a65536f99a770fda8b99"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.429427 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bhktd" podStartSLOduration=134.429403154 podStartE2EDuration="2m14.429403154s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.426476869 +0000 UTC m=+157.667235408" watchObservedRunningTime="2025-12-09 11:33:52.429403154 +0000 UTC m=+157.670161673" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.443231 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.943184248 +0000 UTC m=+158.183942767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.443329 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.444572 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.445189 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:52.945175499 +0000 UTC m=+158.185934018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.460297 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9xgcj" podStartSLOduration=13.460270706 podStartE2EDuration="13.460270706s" podCreationTimestamp="2025-12-09 11:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.449215063 +0000 UTC m=+157.689973582" watchObservedRunningTime="2025-12-09 11:33:52.460270706 +0000 UTC m=+157.701029225" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.482382 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" event={"ID":"c891628b-1ab0-498d-8792-60029bbff6b3","Type":"ContainerStarted","Data":"b50485eb4acb26d50303b023cc994974b12588b38fdc42e0aa563cf17156c2b9"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.483377 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.508456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" event={"ID":"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f","Type":"ContainerStarted","Data":"0e3e6a8e0c829999802fedb626eb466ce74bf3d1effde6a72f83473af697eab3"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.529514 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podStartSLOduration=134.529485281 podStartE2EDuration="2m14.529485281s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.523777795 +0000 UTC m=+157.764536324" watchObservedRunningTime="2025-12-09 11:33:52.529485281 +0000 UTC m=+157.770243810" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" event={"ID":"6684f72f-acbf-4c83-a6d6-1aa9e32a9941","Type":"ContainerStarted","Data":"9e887e97aa12c23f158032b2ca69a0e80e55e4887c4be303f4225e02142a1572"} Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537261 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7rg7c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537365 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gr925 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537399 4770 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d8ftw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537436 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537454 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podUID="3dc28a15-7069-4b05-aced-372d68a397ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.537302 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podUID="6d970f00-0a9d-4259-b21a-503e1e0bceb2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.545467 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.547893 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.047853172 +0000 UTC m=+158.288611741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.552854 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.554139 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.054121583 +0000 UTC m=+158.294880162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.569320 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" podStartSLOduration=134.569302853 podStartE2EDuration="2m14.569302853s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.567565508 +0000 UTC m=+157.808324027" watchObservedRunningTime="2025-12-09 11:33:52.569302853 +0000 UTC m=+157.810061372" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.599806 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-64jcp" podStartSLOduration=134.599788284 podStartE2EDuration="2m14.599788284s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:52.597597598 +0000 UTC m=+157.838356147" watchObservedRunningTime="2025-12-09 11:33:52.599788284 +0000 UTC m=+157.840546803" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.654158 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.654577 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.154561669 +0000 UTC m=+158.395320188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.697794 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cz9jk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.697864 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" podUID="a58a8c2f-222c-41c7-8189-2610471f69df" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.697800 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cz9jk container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.698274 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" podUID="a58a8c2f-222c-41c7-8189-2610471f69df" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.756607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.757155 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.25713978 +0000 UTC m=+158.497898299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.858406 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.858988 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.358962222 +0000 UTC m=+158.599720751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:52 crc kubenswrapper[4770]: I1209 11:33:52.977728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:52 crc kubenswrapper[4770]: E1209 11:33:52.978074 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.478059977 +0000 UTC m=+158.718818486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:52.994832 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7rg7c container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:52.995451 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podUID="6d970f00-0a9d-4259-b21a-503e1e0bceb2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:52.996042 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7rg7c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:52.996122 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podUID="6d970f00-0a9d-4259-b21a-503e1e0bceb2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.025043 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gr925 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.025151 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podUID="3dc28a15-7069-4b05-aced-372d68a397ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.025278 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gr925 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.025311 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" podUID="3dc28a15-7069-4b05-aced-372d68a397ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.272209 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.272600 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.772580231 +0000 UTC m=+159.013338750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.273700 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.387252 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.387592 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:53 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:53 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:53 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.387653 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.389516 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:53.889492549 +0000 UTC m=+159.130251068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.578142 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.578472 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.078447386 +0000 UTC m=+159.319205905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.731169 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.732055 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.232040135 +0000 UTC m=+159.472798654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.769599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" event={"ID":"e26ca10a-d6e1-4b64-a7b9-5b4daceae7a9","Type":"ContainerStarted","Data":"6c5e9c7c4ed5ce609b3c84adee05c58ee5b09d96c6f7c71896bbbecd41bd5d94"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.778296 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" event={"ID":"f0fc13ea-940c-43f5-b256-245017e6664f","Type":"ContainerStarted","Data":"1c81957d5531961e3bdbf1e858d0c7cefa9034dfc1ab59af457bf9df86cea1bd"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.817184 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" event={"ID":"12622b53-0ff2-40ac-9e7e-78b3dcaf8e38","Type":"ContainerStarted","Data":"c4e2aefdb394916b4901d33a2d0bc5ba33fe22123d4a4d040003259573e62259"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.832633 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.834007 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.33398753 +0000 UTC m=+159.574746049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.848364 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" event={"ID":"e215f2a5-db5f-4828-8824-bd30bc0f6c4d","Type":"ContainerStarted","Data":"62803737db7a05571cd51a3266f94ccb713ba61daa86c6197d40d0247654e16b"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.916678 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" event={"ID":"45e8e8a9-2925-41c1-b8f5-7aafde0d700d","Type":"ContainerStarted","Data":"a8d9b4610c8b9261182ae52dfb533392938f1fc4b55a79ae65e35a97380c6be1"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.935996 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:53 crc kubenswrapper[4770]: E1209 11:33:53.936843 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.436823668 +0000 UTC m=+159.677582187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.967380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q9hbn" event={"ID":"348109cc-1761-4c83-b4fb-cdf2e5aa7e1a","Type":"ContainerStarted","Data":"871592fe0dea47fa598db8d23e5ffe5d7fa89d04e0d385441e06420b2fbb57bd"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.976227 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" event={"ID":"94d68ac6-09be-4ca6-8221-f508337ec0c7","Type":"ContainerStarted","Data":"7da83339064de66aef2ffeeb28a472af93299d9111a3a0d6ed18c271c2cce8b4"} Dec 09 11:33:53 crc kubenswrapper[4770]: I1209 11:33:53.977400 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:53.994136 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sfxhc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:53.994209 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.093291 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.095066 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.595042565 +0000 UTC m=+159.835801144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.105138 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" event={"ID":"412ed7a7-4ea3-4daf-8978-a70cf1702c1e","Type":"ContainerStarted","Data":"65cd5f4920be9367a497bef3fafb73f9e1166b34a863d820cdac8fb8adce7a3f"} Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.188946 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" event={"ID":"25c0c5e8-78eb-427b-88f9-e02365a6afc1","Type":"ContainerStarted","Data":"b9db40d328d699d3a00e51f7abe496c1631af1fbc6c658d509ded3f6baf2dd54"} Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.219881 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.224818 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.724795274 +0000 UTC m=+159.965553793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.272027 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:54 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:54 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:54 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.272121 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.272291 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.272334 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.366666 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.367131 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.867083783 +0000 UTC m=+160.107842302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.367363 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.368996 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.868973591 +0000 UTC m=+160.109732310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.485928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.498158 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:54.998123554 +0000 UTC m=+160.238882073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.610891 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.611290 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.111271306 +0000 UTC m=+160.352029825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.717291 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.717618 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.217601383 +0000 UTC m=+160.458359902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.819371 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.845270 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.345247447 +0000 UTC m=+160.586005966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.858301 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fsglb" podStartSLOduration=135.858276361 podStartE2EDuration="2m15.858276361s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:54.846254102 +0000 UTC m=+160.087012631" watchObservedRunningTime="2025-12-09 11:33:54.858276361 +0000 UTC m=+160.099034870" Dec 09 11:33:54 crc kubenswrapper[4770]: I1209 11:33:54.920890 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:54 crc kubenswrapper[4770]: E1209 11:33:54.921345 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.421325928 +0000 UTC m=+160.662084447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.041470 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.042744 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.542716532 +0000 UTC m=+160.783475051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.078455 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" podStartSLOduration=137.078430967 podStartE2EDuration="2m17.078430967s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:54.944265386 +0000 UTC m=+160.185023905" watchObservedRunningTime="2025-12-09 11:33:55.078430967 +0000 UTC m=+160.319189486" Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.302690 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.302734 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7rg7c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.302780 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" podUID="6d970f00-0a9d-4259-b21a-503e1e0bceb2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.303204 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.803178461 +0000 UTC m=+161.043936980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.312249 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:55 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:55 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:55 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.312701 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.313872 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" event={"ID":"a3d2442f-1a13-4e9a-903c-6ed4af79ca24","Type":"ContainerStarted","Data":"90bc7bcf36b30cb20fb59dd67d7f17b617015fd23066939fe681cf40d573685c"} Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.314931 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.319739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" event={"ID":"e215f2a5-db5f-4828-8824-bd30bc0f6c4d","Type":"ContainerStarted","Data":"b173b72978d8d6b3736a5b93548bdf27cf821c546ef6bd55d2d03f2fdc76362b"} Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.322538 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" event={"ID":"365d6032-3c70-47f9-ae54-632097bd0a47","Type":"ContainerStarted","Data":"1c6ab53c5296f1f4d369d9beadbe5bd9e2e23f7ee4c3126e69669dbf3870886f"} Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.325183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" event={"ID":"32fdb059-5acb-446d-98ac-8239919a90b7","Type":"ContainerStarted","Data":"9c86903efb500b6532b6410cd61bec6f434afe35a6e46f3cd25af9429cfe0ab5"} Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.325792 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sfxhc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.325847 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.326430 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.326486 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.404474 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.404987 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:55.904962691 +0000 UTC m=+161.145721210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.505921 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.507200 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.007181733 +0000 UTC m=+161.247940252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.608859 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.609357 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.109336653 +0000 UTC m=+161.350095182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.985688 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.985874 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.48584255 +0000 UTC m=+161.726601079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:55 crc kubenswrapper[4770]: I1209 11:33:55.986051 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:55 crc kubenswrapper[4770]: E1209 11:33:55.986835 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.486821575 +0000 UTC m=+161.727580094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.000328 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.000793 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.000336 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.001155 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.030062 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mwlr8" podStartSLOduration=138.030033213 podStartE2EDuration="2m18.030033213s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.015231164 +0000 UTC m=+161.255989693" watchObservedRunningTime="2025-12-09 11:33:56.030033213 +0000 UTC m=+161.270791732" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.032130 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dg9hr" podStartSLOduration=138.032113307 podStartE2EDuration="2m18.032113307s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:55.092543649 +0000 UTC m=+160.333302178" watchObservedRunningTime="2025-12-09 11:33:56.032113307 +0000 UTC m=+161.272871826" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.164289 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.164735 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.164922 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.165462 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.665392425 +0000 UTC m=+161.906150944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.166556 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.166959 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.666940765 +0000 UTC m=+161.907699284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.207834 4770 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-dzk4z container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.207892 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" podUID="25c0c5e8-78eb-427b-88f9-e02365a6afc1" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.248219 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.248517 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.267176 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.267756 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.76772917 +0000 UTC m=+162.008487689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.289660 4770 patch_prober.go:28] interesting pod/apiserver-76f77b778f-j4njp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.289741 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" podUID="e215f2a5-db5f-4828-8824-bd30bc0f6c4d" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.296164 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:56 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:56 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:56 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.296247 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.351222 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" podStartSLOduration=138.35118799 podStartE2EDuration="2m18.35118799s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.238371037 +0000 UTC m=+161.479129556" watchObservedRunningTime="2025-12-09 11:33:56.35118799 +0000 UTC m=+161.591946509" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.351960 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xn9f7" podStartSLOduration=138.35195264 podStartE2EDuration="2m18.35195264s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.335501538 +0000 UTC m=+161.576260067" watchObservedRunningTime="2025-12-09 11:33:56.35195264 +0000 UTC m=+161.592711159" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.373802 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.374236 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.874221191 +0000 UTC m=+162.114979710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.487133 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.487888 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:56.987862466 +0000 UTC m=+162.228620995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.491620 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-q9hbn" podStartSLOduration=17.491597722 podStartE2EDuration="17.491597722s" podCreationTimestamp="2025-12-09 11:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.490098623 +0000 UTC m=+161.730857142" watchObservedRunningTime="2025-12-09 11:33:56.491597722 +0000 UTC m=+161.732356231" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.518125 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" event={"ID":"32fdb059-5acb-446d-98ac-8239919a90b7","Type":"ContainerStarted","Data":"69f4acf1f0aa1e558483f571496f90eec5dc900f11e57fbc614f9892f34c3e9a"} Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.530776 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" event={"ID":"3cd3eef7-d4d9-4f4c-aba7-d445e622383b","Type":"ContainerStarted","Data":"74b9a5f457e284c88395bd75cfdb01f867b98321b245d1f5d62a3e7b830e11eb"} Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.533227 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sfxhc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.533262 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.591587 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.592539 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.09251921 +0000 UTC m=+162.333277729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.729206 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2gwrb" podStartSLOduration=138.729173325 podStartE2EDuration="2m18.729173325s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.61357769 +0000 UTC m=+161.854336209" watchObservedRunningTime="2025-12-09 11:33:56.729173325 +0000 UTC m=+161.969931864" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.765502 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gpl9q" podStartSLOduration=138.765463476 podStartE2EDuration="2m18.765463476s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:56.756151047 +0000 UTC m=+161.996909566" watchObservedRunningTime="2025-12-09 11:33:56.765463476 +0000 UTC m=+162.006221995" Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.766639 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.768954 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.268917745 +0000 UTC m=+162.509676404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:56 crc kubenswrapper[4770]: I1209 11:33:56.868954 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:56 crc kubenswrapper[4770]: E1209 11:33:56.869515 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.369495294 +0000 UTC m=+162.610253813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.086991 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.087546 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.587516496 +0000 UTC m=+162.828275015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.188787 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.189307 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.689289626 +0000 UTC m=+162.930048155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.320435 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-8t95q" podStartSLOduration=139.320414169 podStartE2EDuration="2m19.320414169s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:57.056518791 +0000 UTC m=+162.297277310" watchObservedRunningTime="2025-12-09 11:33:57.320414169 +0000 UTC m=+162.561172688" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.320881 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.321664 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.326764 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.327111 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.82706666 +0000 UTC m=+163.067825179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.327237 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.327267 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.327315 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.327689 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.827669246 +0000 UTC m=+163.068427845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.332552 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.332949 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.347967 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:57 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:57 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:57 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.348028 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.355848 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.480997 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.481348 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.481402 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.482001 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.482254 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:57.982204149 +0000 UTC m=+163.222962668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.583698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.584270 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.084245116 +0000 UTC m=+163.325003625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.760725 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.761412 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.26138306 +0000 UTC m=+163.502141579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.862053 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.863060 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:57 crc kubenswrapper[4770]: E1209 11:33:57.863420 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.363406896 +0000 UTC m=+163.604165415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:57 crc kubenswrapper[4770]: I1209 11:33:57.911039 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" podStartSLOduration=138.911019698 podStartE2EDuration="2m18.911019698s" podCreationTimestamp="2025-12-09 11:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:57.636883886 +0000 UTC m=+162.877642405" watchObservedRunningTime="2025-12-09 11:33:57.911019698 +0000 UTC m=+163.151778217" Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.067704 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.068134 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.068240 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.568216889 +0000 UTC m=+163.808975408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.069256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.069638 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.569622675 +0000 UTC m=+163.810381204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.289494 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.289806 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.789777742 +0000 UTC m=+164.030536261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.289867 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.290274 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.790259414 +0000 UTC m=+164.031017933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.295237 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:58 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:58 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:58 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.295322 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.390843 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.391342 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:58.891325887 +0000 UTC m=+164.132084406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.574385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.575081 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.075062449 +0000 UTC m=+164.315820968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.684645 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.686709 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" event={"ID":"3cd3eef7-d4d9-4f4c-aba7-d445e622383b","Type":"ContainerStarted","Data":"27de0e542fd181115bf23222446601e2aed104685fdcaf61639161fa65f7567d"} Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.687112 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.187078172 +0000 UTC m=+164.427836691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.687308 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.687744 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.187729549 +0000 UTC m=+164.428488068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:58 crc kubenswrapper[4770]: I1209 11:33:58.808087 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:58 crc kubenswrapper[4770]: E1209 11:33:58.810060 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.310042765 +0000 UTC m=+164.550801284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.063259 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:59 crc kubenswrapper[4770]: E1209 11:33:59.063981 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.563954647 +0000 UTC m=+164.804713166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gxm6f" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.064806 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.064993 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.065186 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.065294 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.217794 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:33:59 crc kubenswrapper[4770]: E1209 11:33:59.218520 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 11:33:59.718499701 +0000 UTC m=+164.959258220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.303012 4770 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.338165 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:33:59 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:33:59 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:33:59 crc kubenswrapper[4770]: healthz check failed Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.338292 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.338545 4770 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-09T11:33:59.303041339Z","Handler":null,"Name":""} Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.523313 4770 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.523359 4770 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.535947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.625733 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.625801 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:33:59 crc kubenswrapper[4770]: I1209 11:33:59.901288 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gxm6f\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.078625 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" podStartSLOduration=142.078608831 podStartE2EDuration="2m22.078608831s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:33:59.332634028 +0000 UTC m=+164.573392547" watchObservedRunningTime="2025-12-09 11:34:00.078608831 +0000 UTC m=+165.319367350" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.079248 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.080424 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.178753 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.179999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnwxz\" (UniqueName: \"kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.180207 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.180479 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: W1209 11:34:00.215473 4770 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 09 11:34:00 crc kubenswrapper[4770]: E1209 11:34:00.215523 4770 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.260549 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.270311 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:00 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:00 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:00 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.270373 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.283498 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnwxz\" (UniqueName: \"kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.283581 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.283659 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.284524 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.284606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.393112 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.555597 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnwxz\" (UniqueName: \"kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz\") pod \"certified-operators-84dm9\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.609996 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.622229 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.623603 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95971ae2-d7fe-4259-b041-9b6373d8a51c-metrics-certs\") pod \"network-metrics-daemon-8pmjn\" (UID: \"95971ae2-d7fe-4259-b041-9b6373d8a51c\") " pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.624270 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.627705 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.637302 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.710691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tjhl\" (UniqueName: \"kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.710797 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.710835 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.725996 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.727404 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.741098 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.742432 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.857329 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.857417 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.877978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tjhl\" (UniqueName: \"kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.878067 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.878107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.883799 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8pmjn" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.902550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:00 crc kubenswrapper[4770]: I1209 11:34:00.902982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.020763 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9xgcj" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.024868 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.024928 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgqmq\" (UniqueName: \"kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.024952 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.024982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.025042 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.025070 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhpsf\" (UniqueName: \"kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.059776 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.060365 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.060460 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.060478 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125439 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhpsf\" (UniqueName: \"kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125548 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgqmq\" (UniqueName: \"kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125620 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125642 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.125696 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.126663 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.129137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.129872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.130245 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.703190 4770 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-84dm9" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.703323 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.704468 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:01 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:01 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:01 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.704540 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.750221 4770 patch_prober.go:28] interesting pod/console-f9d7485db-mhgdc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.750303 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mhgdc" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.939236 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.939992 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.940010 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" event={"ID":"3cd3eef7-d4d9-4f4c-aba7-d445e622383b","Type":"ContainerStarted","Data":"103fc524e449e26dfdcfd3b311abd56f34f9c42717432a33327acc3b5b9224f5"} Dec 09 11:34:01 crc kubenswrapper[4770]: I1209 11:34:01.957390 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.028327 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.028402 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.028503 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.028526 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.028566 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.029372 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"b50485eb4acb26d50303b023cc994974b12588b38fdc42e0aa563cf17156c2b9"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.029464 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" containerID="cri-o://b50485eb4acb26d50303b023cc994974b12588b38fdc42e0aa563cf17156c2b9" gracePeriod=30 Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.029824 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.029849 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.040072 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.172061 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tjhl\" (UniqueName: \"kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl\") pod \"community-operators-cgrjv\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.173225 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgqmq\" (UniqueName: \"kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq\") pod \"certified-operators-ktnmb\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.204179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhpsf\" (UniqueName: \"kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf\") pod \"community-operators-6z5lm\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.212274 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.269229 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:02 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:02 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:02 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.269317 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.286978 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.310197 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.312041 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.337452 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.338200 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.341448 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.343082 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.343689 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.349558 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.350923 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.354974 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.361111 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dzk4z" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.368499 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.376472 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.377807 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.377929 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.383194 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.491246 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.491742 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.510102 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.526621 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.537880 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538153 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538220 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538315 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b7cq\" (UniqueName: \"kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538354 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4vc6\" (UniqueName: \"kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538416 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538467 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnsf7\" (UniqueName: \"kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538530 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kwv8\" (UniqueName: \"kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538586 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538621 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.538646 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.652846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b7cq\" (UniqueName: \"kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653493 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4vc6\" (UniqueName: \"kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653688 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnsf7\" (UniqueName: \"kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653724 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kwv8\" (UniqueName: \"kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.653778 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.654652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.655548 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.655944 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.656460 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.657126 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.657526 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.665472 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.665481 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.731475 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cz9jk" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.770954 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b7cq\" (UniqueName: \"kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq\") pod \"redhat-marketplace-fg6m5\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.805325 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnsf7\" (UniqueName: \"kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7\") pod \"redhat-marketplace-vnc6l\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.808344 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.825551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kwv8\" (UniqueName: \"kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8\") pod \"redhat-operators-cng6d\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.827744 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4vc6\" (UniqueName: \"kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6\") pod \"redhat-operators-rpcqb\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.877788 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.933250 4770 generic.go:334] "Generic (PLEG): container finished" podID="0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" containerID="0e3e6a8e0c829999802fedb626eb466ce74bf3d1effde6a72f83473af697eab3" exitCode=0 Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.933913 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" event={"ID":"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f","Type":"ContainerDied","Data":"0e3e6a8e0c829999802fedb626eb466ce74bf3d1effde6a72f83473af697eab3"} Dec 09 11:34:02 crc kubenswrapper[4770]: I1209 11:34:02.968944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fcadae42-8abb-4f07-8802-e30f5e80838b","Type":"ContainerStarted","Data":"2307188b6027ee754556798423af61a543b0f3516b103ba9b1ab723c3d1fb9d7"} Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.008611 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.027757 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7rg7c" Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.037018 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.095159 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gr925" Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.423712 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.456878 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:03 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:03 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:03 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:03 crc kubenswrapper[4770]: I1209 11:34:03.456983 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:04 crc kubenswrapper[4770]: I1209 11:34:04.128372 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fcadae42-8abb-4f07-8802-e30f5e80838b","Type":"ContainerStarted","Data":"85a64e58b4bd5e730add973d2239aad33720678046a57dc03e22671606a2bf94"} Dec 09 11:34:04 crc kubenswrapper[4770]: I1209 11:34:04.334592 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:04 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:04 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:04 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:04 crc kubenswrapper[4770]: I1209 11:34:04.335049 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.083048 4770 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wdt8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.083396 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" podUID="c891628b-1ab0-498d-8792-60029bbff6b3" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.101575 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8pmjn"] Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.197800 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.249308 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-wdt8r_c891628b-1ab0-498d-8792-60029bbff6b3/openshift-config-operator/0.log" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.253296 4770 generic.go:334] "Generic (PLEG): container finished" podID="c891628b-1ab0-498d-8792-60029bbff6b3" containerID="b50485eb4acb26d50303b023cc994974b12588b38fdc42e0aa563cf17156c2b9" exitCode=255 Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.254145 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" event={"ID":"c891628b-1ab0-498d-8792-60029bbff6b3","Type":"ContainerDied","Data":"b50485eb4acb26d50303b023cc994974b12588b38fdc42e0aa563cf17156c2b9"} Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.254177 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" event={"ID":"c891628b-1ab0-498d-8792-60029bbff6b3","Type":"ContainerStarted","Data":"5e08f526ed1754138df7b301e35c79675b7ffab2d93e1d07f48f637090720639"} Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.254629 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.259109 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.260091 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.266568 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.267413 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.279117 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:05 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:05 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:05 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.279180 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.281011 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.296052 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.296100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.308665 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.468522 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.468566 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.469648 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.587260 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vpb8g" podStartSLOduration=26.587242347 podStartE2EDuration="26.587242347s" podCreationTimestamp="2025-12-09 11:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:34:05.567621974 +0000 UTC m=+170.808380493" watchObservedRunningTime="2025-12-09 11:34:05.587242347 +0000 UTC m=+170.828000866" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.588343 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:05 crc kubenswrapper[4770]: I1209 11:34:05.645335 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.035238 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=9.035197316 podStartE2EDuration="9.035197316s" podCreationTimestamp="2025-12-09 11:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:34:05.844122796 +0000 UTC m=+171.084881315" watchObservedRunningTime="2025-12-09 11:34:06.035197316 +0000 UTC m=+171.275955835" Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.050022 4770 patch_prober.go:28] interesting pod/apiserver-76f77b778f-j4njp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]log ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]etcd ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/max-in-flight-filter ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 09 11:34:06 crc kubenswrapper[4770]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/project.openshift.io-projectcache ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-startinformers ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 11:34:06 crc kubenswrapper[4770]: livez check failed Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.050180 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" podUID="e215f2a5-db5f-4828-8824-bd30bc0f6c4d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:06 crc kubenswrapper[4770]: W1209 11:34:06.161283 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92fca08c_4561_4733_a9d2_488e013b0da1.slice/crio-abd162848a0002a3568f5681c29b73cfc2abf2bd93853f3e6326ef2d3f32dc8d WatchSource:0}: Error finding container abd162848a0002a3568f5681c29b73cfc2abf2bd93853f3e6326ef2d3f32dc8d: Status 404 returned error can't find the container with id abd162848a0002a3568f5681c29b73cfc2abf2bd93853f3e6326ef2d3f32dc8d Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.181310 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.270802 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:06 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.271247 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.274348 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.293569 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" event={"ID":"95971ae2-d7fe-4259-b041-9b6373d8a51c","Type":"ContainerStarted","Data":"ef4a680c86468a32f165181f7fb55b0792d9a0caf05cba1f6938b77de20c05e2"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.304742 4770 patch_prober.go:28] interesting pod/apiserver-76f77b778f-j4njp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]log ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]etcd ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/max-in-flight-filter ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 09 11:34:06 crc kubenswrapper[4770]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/project.openshift.io-projectcache ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-startinformers ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 09 11:34:06 crc kubenswrapper[4770]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 11:34:06 crc kubenswrapper[4770]: livez check failed Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.304832 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" podUID="e215f2a5-db5f-4828-8824-bd30bc0f6c4d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.327692 4770 generic.go:334] "Generic (PLEG): container finished" podID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerID="73f827ed94b8021ade3fc948aeb05e0a079cf4cd524778b88262ea81959314e0" exitCode=0 Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.328043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerDied","Data":"73f827ed94b8021ade3fc948aeb05e0a079cf4cd524778b88262ea81959314e0"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.328138 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerStarted","Data":"35f296ed72881f77f4ee6c2272f5f9fea47f8cf1e4508405cdeb52faedd3fceb"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.375200 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerStarted","Data":"abd162848a0002a3568f5681c29b73cfc2abf2bd93853f3e6326ef2d3f32dc8d"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.388615 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.434854 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.453435 4770 generic.go:334] "Generic (PLEG): container finished" podID="fcadae42-8abb-4f07-8802-e30f5e80838b" containerID="85a64e58b4bd5e730add973d2239aad33720678046a57dc03e22671606a2bf94" exitCode=0 Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.453586 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fcadae42-8abb-4f07-8802-e30f5e80838b","Type":"ContainerDied","Data":"85a64e58b4bd5e730add973d2239aad33720678046a57dc03e22671606a2bf94"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.481743 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.504063 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.520974 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.527552 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerStarted","Data":"e46abe21ea5e58e0dbd03e1ba614f27b827ddd2d5a74285f928492ba485e9dba"} Dec 09 11:34:06 crc kubenswrapper[4770]: I1209 11:34:06.527592 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerStarted","Data":"ff7778012e3751cd84152cbf29e5f28b702b885c2d5123d417e4272121bd208d"} Dec 09 11:34:06 crc kubenswrapper[4770]: W1209 11:34:06.644732 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a151ffe_1d66_4d45_bd04_604d30df30a1.slice/crio-d31bdaf35e7b12b5dffc8044d22bc441d382b2abf701c8805455aa9256d7210d WatchSource:0}: Error finding container d31bdaf35e7b12b5dffc8044d22bc441d382b2abf701c8805455aa9256d7210d: Status 404 returned error can't find the container with id d31bdaf35e7b12b5dffc8044d22bc441d382b2abf701c8805455aa9256d7210d Dec 09 11:34:07 crc kubenswrapper[4770]: E1209 11:34:07.008720 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92fca08c_4561_4733_a9d2_488e013b0da1.slice/crio-conmon-1be4e0aa0fa7ea65d247e0e2b46f3bec0abe9155d60b1b5e84138cd34b3dcb47.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92fca08c_4561_4733_a9d2_488e013b0da1.slice/crio-1be4e0aa0fa7ea65d247e0e2b46f3bec0abe9155d60b1b5e84138cd34b3dcb47.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.047868 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.064425 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.179251 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.271637 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:07 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:07 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:07 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.271682 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.355242 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume\") pod \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.355297 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj7sg\" (UniqueName: \"kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg\") pod \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.355369 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume\") pod \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\" (UID: \"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f\") " Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.356956 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" (UID: "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.459250 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.532269 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" (UID: "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.540031 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg" (OuterVolumeSpecName: "kube-api-access-kj7sg") pod "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" (UID: "0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f"). InnerVolumeSpecName "kube-api-access-kj7sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.561227 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.561300 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj7sg\" (UniqueName: \"kubernetes.io/projected/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f-kube-api-access-kj7sg\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.604372 4770 generic.go:334] "Generic (PLEG): container finished" podID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerID="1dddc829d6c14f034293742bc30b0847e27a62076a5e426e755b094f9ac9c287" exitCode=0 Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.604479 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerDied","Data":"1dddc829d6c14f034293742bc30b0847e27a62076a5e426e755b094f9ac9c287"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.604511 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerStarted","Data":"1ba105edc28a244ec5dab91dabac077883dcb7b1a76dec29a587d6a2240ca5e9"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.611568 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" event={"ID":"0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f","Type":"ContainerDied","Data":"05871e2814e0a333a4be273f3b40d37d099e667f3b3661ad2015889c5adf7dc4"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.611616 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05871e2814e0a333a4be273f3b40d37d099e667f3b3661ad2015889c5adf7dc4" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.611698 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9" Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.621677 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerID="e46abe21ea5e58e0dbd03e1ba614f27b827ddd2d5a74285f928492ba485e9dba" exitCode=0 Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.621765 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerDied","Data":"e46abe21ea5e58e0dbd03e1ba614f27b827ddd2d5a74285f928492ba485e9dba"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.625953 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" event={"ID":"bd7906e7-dbf7-4668-aec9-8e9b778c9452","Type":"ContainerStarted","Data":"612e39b1bd3c342112009aec75a72feb29e7afbc75db6bcfb6f80d88b0224291"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.630993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerStarted","Data":"8d2769f08575d7e9211842622a722c686e89396ececad7b244701d03ee62a31b"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.632675 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerStarted","Data":"88ffc329ced6accd5d6af6939b9a086e012da28293b001cc5f112071111a7e76"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.706723 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" event={"ID":"95971ae2-d7fe-4259-b041-9b6373d8a51c","Type":"ContainerStarted","Data":"dc6701f867df1110747e3957fd92fd0fb633ee054d493e5bc65d3daae97632de"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.706775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8pmjn" event={"ID":"95971ae2-d7fe-4259-b041-9b6373d8a51c","Type":"ContainerStarted","Data":"84fae50fc2f73d933b60e894a0abe100fb15f3bff642b8812873178a97a20dec"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.717359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"22e34886-243b-46bb-a976-937d4808b4d6","Type":"ContainerStarted","Data":"52c0689565e3048656350b2f452d06b0c9cb2d717b7ed927b9455fa1bbab77c9"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.740370 4770 generic.go:334] "Generic (PLEG): container finished" podID="92fca08c-4561-4733-a9d2-488e013b0da1" containerID="1be4e0aa0fa7ea65d247e0e2b46f3bec0abe9155d60b1b5e84138cd34b3dcb47" exitCode=0 Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.740504 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerDied","Data":"1be4e0aa0fa7ea65d247e0e2b46f3bec0abe9155d60b1b5e84138cd34b3dcb47"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.763506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerStarted","Data":"ce7966742fe50f6d6375369f29e47625148027df3374b776249200da9001e8f8"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.768381 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerStarted","Data":"d31bdaf35e7b12b5dffc8044d22bc441d382b2abf701c8805455aa9256d7210d"} Dec 09 11:34:07 crc kubenswrapper[4770]: I1209 11:34:07.780832 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8pmjn" podStartSLOduration=149.780805598 podStartE2EDuration="2m29.780805598s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:34:07.733247008 +0000 UTC m=+172.974005527" watchObservedRunningTime="2025-12-09 11:34:07.780805598 +0000 UTC m=+173.021564117" Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.105815 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wdt8r" Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.263459 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:08 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:08 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:08 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.263556 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.783702 4770 generic.go:334] "Generic (PLEG): container finished" podID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerID="fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075" exitCode=0 Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.784043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerDied","Data":"fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.786995 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" event={"ID":"bd7906e7-dbf7-4668-aec9-8e9b778c9452","Type":"ContainerStarted","Data":"96dca8ae8546027ce121a9bf5912cbb03315e67eff86b61f3c65872a42d8f069"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.788534 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.797892 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"22e34886-243b-46bb-a976-937d4808b4d6","Type":"ContainerStarted","Data":"8abee48ea6dda23991eac4651d56fff61f9f945c5ed56f34373b263e8bb5a99a"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.799800 4770 generic.go:334] "Generic (PLEG): container finished" podID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerID="0488e723a757ad78544fe9badf548b399028787eef36d176bb2209c94f8b74d6" exitCode=0 Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.799860 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerDied","Data":"0488e723a757ad78544fe9badf548b399028787eef36d176bb2209c94f8b74d6"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.806550 4770 generic.go:334] "Generic (PLEG): container finished" podID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerID="ec18afe402bdd1f5f778fb6773869b4ad784e8d71a29f441556bb68f7f927bb8" exitCode=0 Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.806689 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerDied","Data":"ec18afe402bdd1f5f778fb6773869b4ad784e8d71a29f441556bb68f7f927bb8"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.815329 4770 generic.go:334] "Generic (PLEG): container finished" podID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerID="8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67" exitCode=0 Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.820730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerDied","Data":"8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67"} Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.863430 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" podStartSLOduration=150.863407215 podStartE2EDuration="2m30.863407215s" podCreationTimestamp="2025-12-09 11:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:34:08.86244193 +0000 UTC m=+174.103200449" watchObservedRunningTime="2025-12-09 11:34:08.863407215 +0000 UTC m=+174.104165734" Dec 09 11:34:08 crc kubenswrapper[4770]: I1209 11:34:08.867966 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.867946171 podStartE2EDuration="3.867946171s" podCreationTimestamp="2025-12-09 11:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:34:08.835963671 +0000 UTC m=+174.076722210" watchObservedRunningTime="2025-12-09 11:34:08.867946171 +0000 UTC m=+174.108704690" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.067610 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.080944 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access\") pod \"fcadae42-8abb-4f07-8802-e30f5e80838b\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.081044 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir\") pod \"fcadae42-8abb-4f07-8802-e30f5e80838b\" (UID: \"fcadae42-8abb-4f07-8802-e30f5e80838b\") " Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.081282 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fcadae42-8abb-4f07-8802-e30f5e80838b" (UID: "fcadae42-8abb-4f07-8802-e30f5e80838b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.152154 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fcadae42-8abb-4f07-8802-e30f5e80838b" (UID: "fcadae42-8abb-4f07-8802-e30f5e80838b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.182990 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcadae42-8abb-4f07-8802-e30f5e80838b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.183025 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcadae42-8abb-4f07-8802-e30f5e80838b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.289304 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:09 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:09 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:09 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:09 crc kubenswrapper[4770]: I1209 11:34:09.289924 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:10 crc kubenswrapper[4770]: I1209 11:34:10.025798 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fcadae42-8abb-4f07-8802-e30f5e80838b","Type":"ContainerDied","Data":"2307188b6027ee754556798423af61a543b0f3516b103ba9b1ab723c3d1fb9d7"} Dec 09 11:34:10 crc kubenswrapper[4770]: I1209 11:34:10.025916 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2307188b6027ee754556798423af61a543b0f3516b103ba9b1ab723c3d1fb9d7" Dec 09 11:34:10 crc kubenswrapper[4770]: I1209 11:34:10.026008 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 11:34:10 crc kubenswrapper[4770]: I1209 11:34:10.261690 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:10 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:10 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:10 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:10 crc kubenswrapper[4770]: I1209 11:34:10.261759 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.057386 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.057474 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.057523 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.057716 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.057791 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.058945 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.059004 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.059218 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602"} pod="openshift-console/downloads-7954f5f757-xwjv4" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.059285 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" containerID="cri-o://1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602" gracePeriod=2 Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.222525 4770 patch_prober.go:28] interesting pod/console-f9d7485db-mhgdc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.222949 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mhgdc" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.342259 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:11 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:11 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:11 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.342363 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.371025 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:34:11 crc kubenswrapper[4770]: I1209 11:34:11.377971 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-j4njp" Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.080631 4770 generic.go:334] "Generic (PLEG): container finished" podID="22e34886-243b-46bb-a976-937d4808b4d6" containerID="8abee48ea6dda23991eac4651d56fff61f9f945c5ed56f34373b263e8bb5a99a" exitCode=0 Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.080730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"22e34886-243b-46bb-a976-937d4808b4d6","Type":"ContainerDied","Data":"8abee48ea6dda23991eac4651d56fff61f9f945c5ed56f34373b263e8bb5a99a"} Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.084451 4770 generic.go:334] "Generic (PLEG): container finished" podID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerID="1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602" exitCode=0 Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.085621 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerDied","Data":"1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602"} Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.342239 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:12 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:12 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:12 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:12 crc kubenswrapper[4770]: I1209 11:34:12.342288 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.118381 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerStarted","Data":"621ac5dd9cc1c8e4abf4a978e0f1a4091430f12ff215387970cd49f7058d8636"} Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.118717 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.118782 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.118815 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.289545 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:13 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:13 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:13 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:13 crc kubenswrapper[4770]: I1209 11:34:13.289633 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.218619 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.218664 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.354724 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:14 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:14 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:14 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.355185 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.459577 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.570378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access\") pod \"22e34886-243b-46bb-a976-937d4808b4d6\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.570429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir\") pod \"22e34886-243b-46bb-a976-937d4808b4d6\" (UID: \"22e34886-243b-46bb-a976-937d4808b4d6\") " Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.570551 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "22e34886-243b-46bb-a976-937d4808b4d6" (UID: "22e34886-243b-46bb-a976-937d4808b4d6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.570835 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e34886-243b-46bb-a976-937d4808b4d6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.575771 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "22e34886-243b-46bb-a976-937d4808b4d6" (UID: "22e34886-243b-46bb-a976-937d4808b4d6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:34:14 crc kubenswrapper[4770]: I1209 11:34:14.694691 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e34886-243b-46bb-a976-937d4808b4d6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.247350 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.257176 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"22e34886-243b-46bb-a976-937d4808b4d6","Type":"ContainerDied","Data":"52c0689565e3048656350b2f452d06b0c9cb2d717b7ed927b9455fa1bbab77c9"} Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.257240 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52c0689565e3048656350b2f452d06b0c9cb2d717b7ed927b9455fa1bbab77c9" Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.257652 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.257744 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.281270 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:15 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:15 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:15 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:15 crc kubenswrapper[4770]: I1209 11:34:15.281331 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:16 crc kubenswrapper[4770]: I1209 11:34:16.274699 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:16 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:16 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:16 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:16 crc kubenswrapper[4770]: I1209 11:34:16.276123 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:17 crc kubenswrapper[4770]: I1209 11:34:17.261676 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:17 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:17 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:17 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:17 crc kubenswrapper[4770]: I1209 11:34:17.261840 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:18 crc kubenswrapper[4770]: I1209 11:34:18.260664 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:18 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:18 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:18 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:18 crc kubenswrapper[4770]: I1209 11:34:18.260726 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:19 crc kubenswrapper[4770]: I1209 11:34:19.267801 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:19 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:19 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:19 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:19 crc kubenswrapper[4770]: I1209 11:34:19.268007 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:20 crc kubenswrapper[4770]: I1209 11:34:20.261121 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ftbnh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:34:20 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Dec 09 11:34:20 crc kubenswrapper[4770]: [+]process-running ok Dec 09 11:34:20 crc kubenswrapper[4770]: healthz check failed Dec 09 11:34:20 crc kubenswrapper[4770]: I1209 11:34:20.261200 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ftbnh" podUID="e42874d1-4f9d-4d9c-9b45-7e3fdaabfd84" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.055463 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.055539 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.055463 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.055623 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.220974 4770 patch_prober.go:28] interesting pod/console-f9d7485db-mhgdc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.221032 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mhgdc" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.261871 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.264048 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ftbnh" Dec 09 11:34:21 crc kubenswrapper[4770]: I1209 11:34:21.600495 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 11:34:30 crc kubenswrapper[4770]: I1209 11:34:30.398140 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.096366 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.096458 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.096890 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.097000 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.699560 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:34:31 crc kubenswrapper[4770]: I1209 11:34:31.709656 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:34:32 crc kubenswrapper[4770]: I1209 11:34:32.473499 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:34:32 crc kubenswrapper[4770]: I1209 11:34:32.473612 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:34:32 crc kubenswrapper[4770]: I1209 11:34:32.879985 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lqkzj" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.190127 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 11:34:34 crc kubenswrapper[4770]: E1209 11:34:34.191468 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcadae42-8abb-4f07-8802-e30f5e80838b" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191491 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcadae42-8abb-4f07-8802-e30f5e80838b" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: E1209 11:34:34.191510 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" containerName="collect-profiles" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191518 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" containerName="collect-profiles" Dec 09 11:34:34 crc kubenswrapper[4770]: E1209 11:34:34.191530 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e34886-243b-46bb-a976-937d4808b4d6" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191537 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e34886-243b-46bb-a976-937d4808b4d6" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191668 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" containerName="collect-profiles" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191682 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcadae42-8abb-4f07-8802-e30f5e80838b" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.191691 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e34886-243b-46bb-a976-937d4808b4d6" containerName="pruner" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.192240 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.195257 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.195643 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.211337 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.391308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.391372 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.492112 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.492188 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.492322 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.514305 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:34 crc kubenswrapper[4770]: I1209 11:34:34.543540 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.193422 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.195297 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.213633 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.290337 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.290409 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.290447 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.391540 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.391607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.391636 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.391727 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.391726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.414181 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access\") pod \"installer-9-crc\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:39 crc kubenswrapper[4770]: I1209 11:34:39.536412 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.059863 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.059927 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.061178 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.061296 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.061420 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.062312 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.062372 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.062551 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"621ac5dd9cc1c8e4abf4a978e0f1a4091430f12ff215387970cd49f7058d8636"} pod="openshift-console/downloads-7954f5f757-xwjv4" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 09 11:34:41 crc kubenswrapper[4770]: I1209 11:34:41.062607 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" containerID="cri-o://621ac5dd9cc1c8e4abf4a978e0f1a4091430f12ff215387970cd49f7058d8636" gracePeriod=2 Dec 09 11:34:43 crc kubenswrapper[4770]: I1209 11:34:43.995387 4770 generic.go:334] "Generic (PLEG): container finished" podID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerID="621ac5dd9cc1c8e4abf4a978e0f1a4091430f12ff215387970cd49f7058d8636" exitCode=0 Dec 09 11:34:43 crc kubenswrapper[4770]: I1209 11:34:43.995466 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerDied","Data":"621ac5dd9cc1c8e4abf4a978e0f1a4091430f12ff215387970cd49f7058d8636"} Dec 09 11:34:43 crc kubenswrapper[4770]: I1209 11:34:43.995714 4770 scope.go:117] "RemoveContainer" containerID="1ea605084e2d93001d064fc1e3de172dd92e7a0cfdeeea2537924cf5b52ea602" Dec 09 11:34:51 crc kubenswrapper[4770]: I1209 11:34:51.055758 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:34:51 crc kubenswrapper[4770]: I1209 11:34:51.056398 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:00 crc kubenswrapper[4770]: E1209 11:35:00.527940 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 11:35:00 crc kubenswrapper[4770]: E1209 11:35:00.528667 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kwv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cng6d_openshift-marketplace(54352a0d-9e92-431b-9f78-6e9a721c760e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:00 crc kubenswrapper[4770]: E1209 11:35:00.530126 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cng6d" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" Dec 09 11:35:01 crc kubenswrapper[4770]: I1209 11:35:01.055112 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:35:01 crc kubenswrapper[4770]: I1209 11:35:01.055240 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:01 crc kubenswrapper[4770]: E1209 11:35:01.772485 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cng6d" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" Dec 09 11:35:01 crc kubenswrapper[4770]: E1209 11:35:01.835766 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 11:35:01 crc kubenswrapper[4770]: E1209 11:35:01.835981 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b7cq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fg6m5_openshift-marketplace(0a151ffe-1d66-4d45-bd04-604d30df30a1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:01 crc kubenswrapper[4770]: E1209 11:35:01.837336 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fg6m5" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" Dec 09 11:35:02 crc kubenswrapper[4770]: I1209 11:35:02.474092 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:35:02 crc kubenswrapper[4770]: I1209 11:35:02.474167 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:35:02 crc kubenswrapper[4770]: I1209 11:35:02.474211 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:35:02 crc kubenswrapper[4770]: I1209 11:35:02.474841 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:35:02 crc kubenswrapper[4770]: I1209 11:35:02.474923 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29" gracePeriod=600 Dec 09 11:35:03 crc kubenswrapper[4770]: I1209 11:35:03.100715 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29" exitCode=0 Dec 09 11:35:03 crc kubenswrapper[4770]: I1209 11:35:03.100842 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29"} Dec 09 11:35:03 crc kubenswrapper[4770]: E1209 11:35:03.574889 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fg6m5" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" Dec 09 11:35:03 crc kubenswrapper[4770]: E1209 11:35:03.650439 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 11:35:03 crc kubenswrapper[4770]: E1209 11:35:03.651191 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qhpsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6z5lm_openshift-marketplace(7ffafda4-fd6e-4831-a20e-f50e5232a3ba): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:03 crc kubenswrapper[4770]: E1209 11:35:03.652361 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6z5lm" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" Dec 09 11:35:05 crc kubenswrapper[4770]: E1209 11:35:05.876197 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6z5lm" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" Dec 09 11:35:05 crc kubenswrapper[4770]: E1209 11:35:05.972840 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 11:35:05 crc kubenswrapper[4770]: E1209 11:35:05.973020 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dgqmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ktnmb_openshift-marketplace(92fca08c-4561-4733-a9d2-488e013b0da1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:05 crc kubenswrapper[4770]: E1209 11:35:05.975147 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ktnmb" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.017953 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.018141 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnsf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vnc6l_openshift-marketplace(df7cd067-cdd9-42c3-a3eb-770391bd9250): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.020256 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vnc6l" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.039130 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.039463 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6tjhl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgrjv_openshift-marketplace(03311274-e8f5-4e5c-8dce-dce3e7235bfb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.040853 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cgrjv" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.044843 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.045017 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h4vc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rpcqb_openshift-marketplace(357916a9-0349-4b3c-b3c2-58af2ee26540): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.046819 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rpcqb" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.071728 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.072462 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nnwxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-84dm9_openshift-marketplace(b9afd8a5-f75a-4aa5-8cba-a4699b25b847): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.073735 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-84dm9" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.147770 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgrjv" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.148286 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-84dm9" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.150046 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vnc6l" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.150156 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rpcqb" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" Dec 09 11:35:06 crc kubenswrapper[4770]: E1209 11:35:06.150550 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ktnmb" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" Dec 09 11:35:06 crc kubenswrapper[4770]: I1209 11:35:06.617598 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 11:35:06 crc kubenswrapper[4770]: I1209 11:35:06.634439 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 11:35:06 crc kubenswrapper[4770]: W1209 11:35:06.653647 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod33248723_2e19_46c6_b13f_73d7a9afded0.slice/crio-dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971 WatchSource:0}: Error finding container dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971: Status 404 returned error can't find the container with id dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971 Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.157260 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.161443 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7","Type":"ContainerStarted","Data":"98004def3e17854951274b8db1394950ce186bf72ca609f702ab59950dd05385"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.161572 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7","Type":"ContainerStarted","Data":"774d786979dd23c2d27490e57ad5bd31bc61d040ffded869763f4559316c6fc0"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.167490 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"33248723-2e19-46c6-b13f-73d7a9afded0","Type":"ContainerStarted","Data":"886c60dba3d044bf73e637505775a4a1e3f4a15ff67683580b45942d3a69b4ab"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.167583 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"33248723-2e19-46c6-b13f-73d7a9afded0","Type":"ContainerStarted","Data":"dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.171393 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xwjv4" event={"ID":"e69f8571-c321-4ff0-9fd1-fb67b1176230","Type":"ContainerStarted","Data":"5a1ee866644520fcf7ab9735494ee12e31b5948cd9d558bce9afa926d04c70fb"} Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.171887 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.172690 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.172776 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:07 crc kubenswrapper[4770]: I1209 11:35:07.263016 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=28.262980654 podStartE2EDuration="28.262980654s" podCreationTimestamp="2025-12-09 11:34:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:35:07.241434258 +0000 UTC m=+232.482192787" watchObservedRunningTime="2025-12-09 11:35:07.262980654 +0000 UTC m=+232.503739183" Dec 09 11:35:08 crc kubenswrapper[4770]: I1209 11:35:08.179824 4770 generic.go:334] "Generic (PLEG): container finished" podID="33248723-2e19-46c6-b13f-73d7a9afded0" containerID="886c60dba3d044bf73e637505775a4a1e3f4a15ff67683580b45942d3a69b4ab" exitCode=0 Dec 09 11:35:08 crc kubenswrapper[4770]: I1209 11:35:08.179942 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"33248723-2e19-46c6-b13f-73d7a9afded0","Type":"ContainerDied","Data":"886c60dba3d044bf73e637505775a4a1e3f4a15ff67683580b45942d3a69b4ab"} Dec 09 11:35:08 crc kubenswrapper[4770]: I1209 11:35:08.182996 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:35:08 crc kubenswrapper[4770]: I1209 11:35:08.183084 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.460618 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.528412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir\") pod \"33248723-2e19-46c6-b13f-73d7a9afded0\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.528542 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access\") pod \"33248723-2e19-46c6-b13f-73d7a9afded0\" (UID: \"33248723-2e19-46c6-b13f-73d7a9afded0\") " Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.528549 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "33248723-2e19-46c6-b13f-73d7a9afded0" (UID: "33248723-2e19-46c6-b13f-73d7a9afded0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.529006 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/33248723-2e19-46c6-b13f-73d7a9afded0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.536495 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "33248723-2e19-46c6-b13f-73d7a9afded0" (UID: "33248723-2e19-46c6-b13f-73d7a9afded0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:35:09 crc kubenswrapper[4770]: I1209 11:35:09.630660 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33248723-2e19-46c6-b13f-73d7a9afded0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:10 crc kubenswrapper[4770]: I1209 11:35:10.202613 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"33248723-2e19-46c6-b13f-73d7a9afded0","Type":"ContainerDied","Data":"dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971"} Dec 09 11:35:10 crc kubenswrapper[4770]: I1209 11:35:10.202662 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd72487a9fa2034b745ea28f8a5f8d1959429acd0e621d076e96d63f69342971" Dec 09 11:35:10 crc kubenswrapper[4770]: I1209 11:35:10.202687 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 11:35:11 crc kubenswrapper[4770]: I1209 11:35:11.055072 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:35:11 crc kubenswrapper[4770]: I1209 11:35:11.055108 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-xwjv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 09 11:35:11 crc kubenswrapper[4770]: I1209 11:35:11.055145 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:11 crc kubenswrapper[4770]: I1209 11:35:11.055165 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xwjv4" podUID="e69f8571-c321-4ff0-9fd1-fb67b1176230" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 09 11:35:21 crc kubenswrapper[4770]: I1209 11:35:21.077331 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xwjv4" Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.447789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerStarted","Data":"9a3f9a6d0bcdb2263b56bb42d75760009e85e9f2a219f015a4d43d28aa1b74f9"} Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.454236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerStarted","Data":"e9ee728d3985ba497c6667c0eb76c6d97f150138f447a02a8d43112874eac6db"} Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.456391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerStarted","Data":"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9"} Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.457643 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerStarted","Data":"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991"} Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.460751 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerStarted","Data":"b3ba907dea74f1a3f20b73a5b54bc7ac0e87a5b711e183d72bc4e9d3dff3c97a"} Dec 09 11:35:36 crc kubenswrapper[4770]: I1209 11:35:36.466003 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerStarted","Data":"b3d670f2a8db28e2d2ed26e3ca1e3a55cf7e179cc66f51b44f0fa7b982088190"} Dec 09 11:35:37 crc kubenswrapper[4770]: I1209 11:35:37.482122 4770 generic.go:334] "Generic (PLEG): container finished" podID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerID="09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991" exitCode=0 Dec 09 11:35:37 crc kubenswrapper[4770]: I1209 11:35:37.482215 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerDied","Data":"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991"} Dec 09 11:35:37 crc kubenswrapper[4770]: I1209 11:35:37.489506 4770 generic.go:334] "Generic (PLEG): container finished" podID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerID="9a3f9a6d0bcdb2263b56bb42d75760009e85e9f2a219f015a4d43d28aa1b74f9" exitCode=0 Dec 09 11:35:37 crc kubenswrapper[4770]: I1209 11:35:37.489649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerDied","Data":"9a3f9a6d0bcdb2263b56bb42d75760009e85e9f2a219f015a4d43d28aa1b74f9"} Dec 09 11:35:38 crc kubenswrapper[4770]: I1209 11:35:38.634848 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerStarted","Data":"39931cd47ae9333df0b2c341c3727f2459ff21f87e4a4c9456c0b6bdc97a3f79"} Dec 09 11:35:38 crc kubenswrapper[4770]: I1209 11:35:38.638384 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerStarted","Data":"639175d28e23ccd2f22f5209a68cb92e6624f9dbfebe1e590ad2d9432db5edef"} Dec 09 11:35:38 crc kubenswrapper[4770]: I1209 11:35:38.643211 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerID="b3d670f2a8db28e2d2ed26e3ca1e3a55cf7e179cc66f51b44f0fa7b982088190" exitCode=0 Dec 09 11:35:38 crc kubenswrapper[4770]: I1209 11:35:38.643284 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerDied","Data":"b3d670f2a8db28e2d2ed26e3ca1e3a55cf7e179cc66f51b44f0fa7b982088190"} Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.649367 4770 generic.go:334] "Generic (PLEG): container finished" podID="92fca08c-4561-4733-a9d2-488e013b0da1" containerID="e9ee728d3985ba497c6667c0eb76c6d97f150138f447a02a8d43112874eac6db" exitCode=0 Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.649848 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerDied","Data":"e9ee728d3985ba497c6667c0eb76c6d97f150138f447a02a8d43112874eac6db"} Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.653699 4770 generic.go:334] "Generic (PLEG): container finished" podID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerID="66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9" exitCode=0 Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.653763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerDied","Data":"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9"} Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.656169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerStarted","Data":"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38"} Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.658039 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerStarted","Data":"32b37d1cd3dc61b78afc9fd2df735d736e635d9faa7feabada5417cb9180034c"} Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.773793 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vnc6l" podStartSLOduration=8.571756972 podStartE2EDuration="1m37.773763505s" podCreationTimestamp="2025-12-09 11:34:02 +0000 UTC" firstStartedPulling="2025-12-09 11:34:08.982068748 +0000 UTC m=+174.222827267" lastFinishedPulling="2025-12-09 11:35:38.184075281 +0000 UTC m=+263.424833800" observedRunningTime="2025-12-09 11:35:39.770612534 +0000 UTC m=+265.011371053" watchObservedRunningTime="2025-12-09 11:35:39.773763505 +0000 UTC m=+265.014522014" Dec 09 11:35:39 crc kubenswrapper[4770]: I1209 11:35:39.870017 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fg6m5" podStartSLOduration=8.735658855 podStartE2EDuration="1m37.869989646s" podCreationTimestamp="2025-12-09 11:34:02 +0000 UTC" firstStartedPulling="2025-12-09 11:34:08.981403611 +0000 UTC m=+174.222162130" lastFinishedPulling="2025-12-09 11:35:38.115734402 +0000 UTC m=+263.356492921" observedRunningTime="2025-12-09 11:35:39.865320845 +0000 UTC m=+265.106079364" watchObservedRunningTime="2025-12-09 11:35:39.869989646 +0000 UTC m=+265.110748165" Dec 09 11:35:41 crc kubenswrapper[4770]: I1209 11:35:41.671506 4770 generic.go:334] "Generic (PLEG): container finished" podID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerID="39931cd47ae9333df0b2c341c3727f2459ff21f87e4a4c9456c0b6bdc97a3f79" exitCode=0 Dec 09 11:35:41 crc kubenswrapper[4770]: I1209 11:35:41.671977 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerDied","Data":"39931cd47ae9333df0b2c341c3727f2459ff21f87e4a4c9456c0b6bdc97a3f79"} Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.642507 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wdxh"] Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.704426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerStarted","Data":"9186eca8ab5b6b6f088b79a4bedaf5039e7602d97c2cfeb8e736335bf0ff68f9"} Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.716169 4770 generic.go:334] "Generic (PLEG): container finished" podID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerID="b3ba907dea74f1a3f20b73a5b54bc7ac0e87a5b711e183d72bc4e9d3dff3c97a" exitCode=0 Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.716239 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerDied","Data":"b3ba907dea74f1a3f20b73a5b54bc7ac0e87a5b711e183d72bc4e9d3dff3c97a"} Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.752352 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84dm9" podStartSLOduration=10.921970935000001 podStartE2EDuration="1m43.752328748s" podCreationTimestamp="2025-12-09 11:33:59 +0000 UTC" firstStartedPulling="2025-12-09 11:34:06.550196735 +0000 UTC m=+171.790955254" lastFinishedPulling="2025-12-09 11:35:39.380554548 +0000 UTC m=+264.621313067" observedRunningTime="2025-12-09 11:35:42.749227758 +0000 UTC m=+267.989986277" watchObservedRunningTime="2025-12-09 11:35:42.752328748 +0000 UTC m=+267.993087267" Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.811461 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.811542 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.879177 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:35:42 crc kubenswrapper[4770]: I1209 11:35:42.879268 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.570035 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-fg6m5" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="registry-server" probeResult="failure" output=< Dec 09 11:35:44 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 11:35:44 crc kubenswrapper[4770]: > Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.579453 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-vnc6l" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="registry-server" probeResult="failure" output=< Dec 09 11:35:44 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 11:35:44 crc kubenswrapper[4770]: > Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.730947 4770 generic.go:334] "Generic (PLEG): container finished" podID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerID="639175d28e23ccd2f22f5209a68cb92e6624f9dbfebe1e590ad2d9432db5edef" exitCode=0 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.731007 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerDied","Data":"639175d28e23ccd2f22f5209a68cb92e6624f9dbfebe1e590ad2d9432db5edef"} Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.832662 4770 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.832924 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33248723-2e19-46c6-b13f-73d7a9afded0" containerName="pruner" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.832936 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="33248723-2e19-46c6-b13f-73d7a9afded0" containerName="pruner" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.833049 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="33248723-2e19-46c6-b13f-73d7a9afded0" containerName="pruner" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.833464 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.834586 4770 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.834980 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8" gracePeriod=15 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.835220 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4" gracePeriod=15 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.835119 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835" gracePeriod=15 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.835185 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd" gracePeriod=15 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.835168 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5" gracePeriod=15 Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.862821 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.973638 4770 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.975258 4770 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.975582 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.975655 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.975718 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.975777 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.975833 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.975887 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.975990 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.976047 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.976103 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.976152 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.976203 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.976258 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: E1209 11:35:44.976342 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.976407 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.977489 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.977743 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.977963 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.978157 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.978260 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 11:35:44 crc kubenswrapper[4770]: I1209 11:35:44.978356 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.004754 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.004866 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.004962 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.005001 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.005074 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.105959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106019 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106047 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106088 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106117 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106135 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106144 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106113 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106156 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106165 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.106342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.159207 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: E1209 11:35:45.190364 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f88f06f131520 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,LastTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207250 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207301 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207354 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207355 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207388 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.207428 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.343542 4770 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.344007 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.738463 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f9d44db2db60f9a934f0568aaed15fa73a011eada6b4b4b86ca428500bcb61e8"} Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.741304 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.743835 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.744980 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835" exitCode=0 Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.745014 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4" exitCode=2 Dec 09 11:35:45 crc kubenswrapper[4770]: I1209 11:35:45.745080 4770 scope.go:117] "RemoveContainer" containerID="9305cd7c4202d58e5e026ae4361f1c08f37e5d927aa7eed46d2f47ea09e9b4c9" Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.753267 4770 generic.go:334] "Generic (PLEG): container finished" podID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" containerID="98004def3e17854951274b8db1394950ce186bf72ca609f702ab59950dd05385" exitCode=0 Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.753397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7","Type":"ContainerDied","Data":"98004def3e17854951274b8db1394950ce186bf72ca609f702ab59950dd05385"} Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.754339 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.754821 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.757491 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.758186 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5" exitCode=0 Dec 09 11:35:46 crc kubenswrapper[4770]: I1209 11:35:46.758208 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd" exitCode=0 Dec 09 11:35:49 crc kubenswrapper[4770]: E1209 11:35:49.499438 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f88f06f131520 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,LastTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 11:35:50 crc kubenswrapper[4770]: I1209 11:35:50.797393 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cccd9cced28043456f1ddb11b6a254e0fd5e85cd65a4ddef5637118a773dd633"} Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.262975 4770 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.263655 4770 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.264175 4770 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.264386 4770 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.264779 4770 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.265080 4770 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.265507 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.325622 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.326411 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.327403 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.396514 4770 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" volumeName="registry-storage" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.407842 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access\") pod \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.408243 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock\") pod \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.408291 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir\") pod \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\" (UID: \"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7\") " Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.408581 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" (UID: "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.408623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock" (OuterVolumeSpecName: "var-lock") pod "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" (UID: "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.416475 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" (UID: "e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.466884 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.509708 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.509763 4770 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.509776 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.706522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.706611 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.807911 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.807969 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7","Type":"ContainerDied","Data":"774d786979dd23c2d27490e57ad5bd31bc61d040ffded869763f4559316c6fc0"} Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.808441 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="774d786979dd23c2d27490e57ad5bd31bc61d040ffded869763f4559316c6fc0" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.811539 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.812952 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8" exitCode=0 Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.814247 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.814983 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.837134 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.837258 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.838023 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.838823 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.839421 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.839859 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: E1209 11:35:51.868340 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.877308 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.878019 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.878551 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:51 crc kubenswrapper[4770]: I1209 11:35:51.879126 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.669524 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.816987 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:35:52Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:35:52Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:35:52Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T11:35:52Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.817721 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.818175 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.818568 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.819045 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: E1209 11:35:52.819071 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.862790 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.863508 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.863751 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.864020 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.864192 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.907629 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.908517 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.909158 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.909525 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.909758 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.922399 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.923117 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.923468 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.923695 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.923934 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.924164 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.965437 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.966229 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.966868 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.967488 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.968066 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:52 crc kubenswrapper[4770]: I1209 11:35:52.968457 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:54 crc kubenswrapper[4770]: E1209 11:35:54.271155 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="3.2s" Dec 09 11:35:55 crc kubenswrapper[4770]: I1209 11:35:55.344293 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:55 crc kubenswrapper[4770]: I1209 11:35:55.344985 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:55 crc kubenswrapper[4770]: I1209 11:35:55.345304 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:55 crc kubenswrapper[4770]: I1209 11:35:55.345690 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:55 crc kubenswrapper[4770]: I1209 11:35:55.346189 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.452298 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.453527 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.454527 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.454875 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.455466 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.455748 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.456021 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.456548 4770 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484272 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484315 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484333 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484444 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484515 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.484504 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.485004 4770 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.485025 4770 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.485037 4770 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.851306 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.852467 4770 scope.go:117] "RemoveContainer" containerID="8eb246002cfaf2546728434fe513ad759b039a798f422d166160b7d3ce960835" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.852604 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.870314 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.870727 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.871233 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.871642 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.872790 4770 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.873478 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.880271 4770 scope.go:117] "RemoveContainer" containerID="5072cdccc16afa1bc4bbe4c1b93f307c7a3c6e11aa694bdb5add0f4a0e968bc5" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.961180 4770 scope.go:117] "RemoveContainer" containerID="3dd34c445832872d20057c406ed213db495653dfa4901f85435ca60d2f65e4dd" Dec 09 11:35:56 crc kubenswrapper[4770]: I1209 11:35:56.988553 4770 scope.go:117] "RemoveContainer" containerID="10e6f7ae1e659a56973a057b15d3a657b5f718f111e7227d70f46def7095dfb4" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.050197 4770 scope.go:117] "RemoveContainer" containerID="d4f69f8f4249846195f528ebdf30f94f44c319d299d199fc400ef7bf49d2b1c8" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.086243 4770 scope.go:117] "RemoveContainer" containerID="4c9b80ecd3d36634e043ad319537b1cb7c843a7be684414b38296859d0e2eb6a" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.339615 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.341479 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.342094 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.342338 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.342822 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.343490 4770 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.343767 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.350130 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.357175 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.357221 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:35:57 crc kubenswrapper[4770]: E1209 11:35:57.358140 4770 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.359109 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:57 crc kubenswrapper[4770]: W1209 11:35:57.382741 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-356962275d445224cc134efbccd1c69e826044ca43664b0b15de98563cfe1e91 WatchSource:0}: Error finding container 356962275d445224cc134efbccd1c69e826044ca43664b0b15de98563cfe1e91: Status 404 returned error can't find the container with id 356962275d445224cc134efbccd1c69e826044ca43664b0b15de98563cfe1e91 Dec 09 11:35:57 crc kubenswrapper[4770]: E1209 11:35:57.473194 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="6.4s" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.862943 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerStarted","Data":"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.863875 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.864410 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"356962275d445224cc134efbccd1c69e826044ca43664b0b15de98563cfe1e91"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.864481 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.865001 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.865267 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.865541 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.865885 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.868669 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerStarted","Data":"028a69e195d308587c22bd9b31ad909606b0269f7c88819cd17e0a2731e821d3"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.869362 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.869716 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.870278 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.870479 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.870738 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.871115 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.871637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerStarted","Data":"7a0bfea976d7b11910250169c1146afa4ec8aaaa7a9d9baf008fad3ceb4c65a7"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.871822 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.872277 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.872484 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.872677 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.873049 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.873584 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.873805 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.874084 4770 status_manager.go:851] "Failed to get status for pod" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" pod="openshift-marketplace/community-operators-cgrjv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-cgrjv\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.874461 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.875588 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerStarted","Data":"d5525354f02add30b14fcf41c39432ca518eeaca5218eff280dbfed5e8708e02"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.876595 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.876859 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.877062 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.877253 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.877426 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.877672 4770 status_manager.go:851] "Failed to get status for pod" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" pod="openshift-marketplace/community-operators-cgrjv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-cgrjv\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.878032 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.878530 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerStarted","Data":"e94ffe2a3195b5b5a16e2cb8a93faaac3f88681c7e9d1185493e88fb1c78e1d1"} Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.878535 4770 status_manager.go:851] "Failed to get status for pod" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" pod="openshift-marketplace/certified-operators-ktnmb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-ktnmb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.879019 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.879493 4770 status_manager.go:851] "Failed to get status for pod" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" pod="openshift-marketplace/certified-operators-ktnmb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-ktnmb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.879884 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.880796 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.881235 4770 status_manager.go:851] "Failed to get status for pod" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" pod="openshift-marketplace/redhat-operators-rpcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-rpcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.881556 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.882566 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.882868 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.883308 4770 status_manager.go:851] "Failed to get status for pod" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" pod="openshift-marketplace/community-operators-cgrjv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-cgrjv\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.883894 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:57 crc kubenswrapper[4770]: I1209 11:35:57.884190 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:58 crc kubenswrapper[4770]: I1209 11:35:58.886139 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8ef125f5214180a8125743d5038a243d3096b1d593951880e9a3453feac1c54b"} Dec 09 11:35:59 crc kubenswrapper[4770]: E1209 11:35:59.502377 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f88f06f131520 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,LastTimestamp:2025-12-09 11:35:45.18957392 +0000 UTC m=+270.430332439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.898289 4770 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="8ef125f5214180a8125743d5038a243d3096b1d593951880e9a3453feac1c54b" exitCode=0 Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.898452 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"8ef125f5214180a8125743d5038a243d3096b1d593951880e9a3453feac1c54b"} Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.899544 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.899586 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.899710 4770 status_manager.go:851] "Failed to get status for pod" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" pod="openshift-marketplace/community-operators-cgrjv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-cgrjv\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: E1209 11:35:59.900036 4770 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.900450 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.900832 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.901447 4770 status_manager.go:851] "Failed to get status for pod" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" pod="openshift-marketplace/certified-operators-ktnmb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-ktnmb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.902054 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.903249 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.903304 4770 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3" exitCode=1 Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.903349 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3"} Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.904015 4770 scope.go:117] "RemoveContainer" containerID="75259146d4f47dadbfcbb49ebd98ed0c4f1012802cfd9ebda4994dfe294625f3" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.904200 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.904831 4770 status_manager.go:851] "Failed to get status for pod" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" pod="openshift-marketplace/redhat-operators-rpcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-rpcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.905449 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.905961 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.906435 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.907571 4770 status_manager.go:851] "Failed to get status for pod" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" pod="openshift-marketplace/community-operators-6z5lm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6z5lm\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.908075 4770 status_manager.go:851] "Failed to get status for pod" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" pod="openshift-marketplace/redhat-operators-cng6d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cng6d\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.908469 4770 status_manager.go:851] "Failed to get status for pod" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" pod="openshift-marketplace/community-operators-cgrjv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-cgrjv\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.908744 4770 status_manager.go:851] "Failed to get status for pod" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.909133 4770 status_manager.go:851] "Failed to get status for pod" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" pod="openshift-marketplace/redhat-marketplace-fg6m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fg6m5\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.909474 4770 status_manager.go:851] "Failed to get status for pod" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" pod="openshift-marketplace/certified-operators-ktnmb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-ktnmb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.909956 4770 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.910302 4770 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.910611 4770 status_manager.go:851] "Failed to get status for pod" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" pod="openshift-marketplace/redhat-operators-rpcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-rpcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.911016 4770 status_manager.go:851] "Failed to get status for pod" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" pod="openshift-marketplace/certified-operators-84dm9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-84dm9\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:35:59 crc kubenswrapper[4770]: I1209 11:35:59.911303 4770 status_manager.go:851] "Failed to get status for pod" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" pod="openshift-marketplace/redhat-marketplace-vnc6l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-vnc6l\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 09 11:36:01 crc kubenswrapper[4770]: I1209 11:36:01.923812 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 11:36:01 crc kubenswrapper[4770]: I1209 11:36:01.924502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"771a65f74586cb16354fa6c0f897f97eb3ee4a70a5b9461ad4781142d9ccf951"} Dec 09 11:36:01 crc kubenswrapper[4770]: I1209 11:36:01.929141 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"181a41db0e48f1b0744111e36a62e04b2f39038fad3a0c921af770e42981a78d"} Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.288687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.289310 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.339536 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.339619 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.342603 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.359663 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.359761 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.391398 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.423227 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.462433 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.987676 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:02.995248 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.004419 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.010093 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.010127 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.038836 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.038884 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.064872 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.239775 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.240309 4770 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.240489 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.941093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b4d03aa45170fde5d8d50e873133ab7b00b406bb692e0d21ebf3aadb506c0866"} Dec 09 11:36:03 crc kubenswrapper[4770]: I1209 11:36:03.984496 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:36:04 crc kubenswrapper[4770]: I1209 11:36:04.083860 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rpcqb" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="registry-server" probeResult="failure" output=< Dec 09 11:36:04 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 11:36:04 crc kubenswrapper[4770]: > Dec 09 11:36:04 crc kubenswrapper[4770]: I1209 11:36:04.951027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c825abe932eb92a65334767cd57b648a06df3f2c71baef0c2f9c3058c29a79b9"} Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.962401 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8fc910f384a778c427e5db0e7e0e05de918c5cad32252646f8f8e62115bd3140"} Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.962474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9b8a3a9c907c344f70feeee1dc62e542d47e37191a6ec1d82842c91d3b0e6217"} Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.962717 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.962800 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.962839 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:05 crc kubenswrapper[4770]: I1209 11:36:05.974738 4770 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:06 crc kubenswrapper[4770]: I1209 11:36:06.967440 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:06 crc kubenswrapper[4770]: I1209 11:36:06.967820 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.359858 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.359947 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.366813 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.703816 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerName="oauth-openshift" containerID="cri-o://7b15bc9df0119f263b45a27ce813b40eb30d2cc433a08532b16a9372aa80fff2" gracePeriod=15 Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.978252 4770 generic.go:334] "Generic (PLEG): container finished" podID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerID="7b15bc9df0119f263b45a27ce813b40eb30d2cc433a08532b16a9372aa80fff2" exitCode=0 Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.978391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" event={"ID":"e1a5a731-6de5-4cfa-abfd-b68487c0b855","Type":"ContainerDied","Data":"7b15bc9df0119f263b45a27ce813b40eb30d2cc433a08532b16a9372aa80fff2"} Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.978829 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.978852 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:07 crc kubenswrapper[4770]: I1209 11:36:07.984285 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.176753 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.285721 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.285793 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.285879 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.285962 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286070 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286107 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppmbt\" (UniqueName: \"kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286139 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286168 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286201 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286289 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286344 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286419 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.286473 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login\") pod \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\" (UID: \"e1a5a731-6de5-4cfa-abfd-b68487c0b855\") " Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.287587 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.288892 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.290099 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.295387 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.303756 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.304192 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.305377 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt" (OuterVolumeSpecName: "kube-api-access-ppmbt") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "kube-api-access-ppmbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.306667 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.307428 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.317650 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.317955 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.320422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.320813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.321096 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e1a5a731-6de5-4cfa-abfd-b68487c0b855" (UID: "e1a5a731-6de5-4cfa-abfd-b68487c0b855"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388668 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388726 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388749 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388772 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388790 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388808 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388828 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388844 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388858 4770 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1a5a731-6de5-4cfa-abfd-b68487c0b855-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388873 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388888 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppmbt\" (UniqueName: \"kubernetes.io/projected/e1a5a731-6de5-4cfa-abfd-b68487c0b855-kube-api-access-ppmbt\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.388998 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.389014 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.389030 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a5a731-6de5-4cfa-abfd-b68487c0b855-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.654924 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5dde30f3-d43b-4364-8205-81dc2d01e842" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.986774 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.986746 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wdxh" event={"ID":"e1a5a731-6de5-4cfa-abfd-b68487c0b855","Type":"ContainerDied","Data":"2d434e29bbb092e2cff42ef58636ae91041b9528a79dccc3044474b868811b92"} Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.986846 4770 scope.go:117] "RemoveContainer" containerID="7b15bc9df0119f263b45a27ce813b40eb30d2cc433a08532b16a9372aa80fff2" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.987102 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.987131 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:08 crc kubenswrapper[4770]: I1209 11:36:08.991471 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5dde30f3-d43b-4364-8205-81dc2d01e842" Dec 09 11:36:09 crc kubenswrapper[4770]: I1209 11:36:09.995474 4770 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:09 crc kubenswrapper[4770]: I1209 11:36:09.995977 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0ff5f962-1825-4fc0-881d-dd3f982e99a2" Dec 09 11:36:10 crc kubenswrapper[4770]: I1209 11:36:10.000401 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5dde30f3-d43b-4364-8205-81dc2d01e842" Dec 09 11:36:13 crc kubenswrapper[4770]: I1209 11:36:13.084673 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:36:13 crc kubenswrapper[4770]: I1209 11:36:13.130497 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:36:13 crc kubenswrapper[4770]: I1209 11:36:13.240996 4770 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 11:36:13 crc kubenswrapper[4770]: I1209 11:36:13.241092 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 11:36:15 crc kubenswrapper[4770]: I1209 11:36:15.194028 4770 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 09 11:36:18 crc kubenswrapper[4770]: I1209 11:36:18.010241 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 11:36:18 crc kubenswrapper[4770]: I1209 11:36:18.909145 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.150936 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.611714 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.765045 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.783601 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.818538 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 11:36:19 crc kubenswrapper[4770]: I1209 11:36:19.911515 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 11:36:20 crc kubenswrapper[4770]: I1209 11:36:20.093858 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 11:36:20 crc kubenswrapper[4770]: I1209 11:36:20.796039 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.218701 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.262598 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.291479 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.376785 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.419264 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.485263 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.610534 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.657645 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.784130 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.836097 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.836100 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.839738 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.849211 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.862387 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.876739 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.876977 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 11:36:21 crc kubenswrapper[4770]: I1209 11:36:21.928746 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.058824 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.102593 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.164737 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.217684 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.222599 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.261674 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.274537 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.504384 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.553018 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.637208 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.726702 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.794079 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.810042 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.892035 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.898602 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.925036 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 11:36:22 crc kubenswrapper[4770]: I1209 11:36:22.935933 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.005941 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.037609 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.069267 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.101517 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.211873 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.245483 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.250691 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.253103 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.286444 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.371938 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.523167 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.603419 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.753780 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.771893 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.842943 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.844742 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 11:36:23 crc kubenswrapper[4770]: I1209 11:36:23.943231 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.017988 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.153588 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.192832 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.239119 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.247089 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.248085 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.251664 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.298338 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.311163 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.360458 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.457397 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.528495 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.544826 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.675920 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.683180 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.712364 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.737046 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.780834 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.800269 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.895256 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.901087 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.942003 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.972867 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 11:36:24 crc kubenswrapper[4770]: I1209 11:36:24.994025 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.131165 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.177117 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.188911 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.259194 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.300930 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.321457 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.386043 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.412677 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.446315 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.475122 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.577331 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.612185 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.626782 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.680061 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.889194 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 11:36:25 crc kubenswrapper[4770]: I1209 11:36:25.894577 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.158683 4770 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.158884 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ktnmb" podStartSLOduration=37.246210829 podStartE2EDuration="2m26.15882106s" podCreationTimestamp="2025-12-09 11:34:00 +0000 UTC" firstStartedPulling="2025-12-09 11:34:07.745001079 +0000 UTC m=+172.985759598" lastFinishedPulling="2025-12-09 11:35:56.6576113 +0000 UTC m=+281.898369829" observedRunningTime="2025-12-09 11:36:08.452836827 +0000 UTC m=+293.693595346" watchObservedRunningTime="2025-12-09 11:36:26.15882106 +0000 UTC m=+311.399579589" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.161961 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6z5lm" podStartSLOduration=38.35935649 podStartE2EDuration="2m26.16195052s" podCreationTimestamp="2025-12-09 11:34:00 +0000 UTC" firstStartedPulling="2025-12-09 11:34:08.979171684 +0000 UTC m=+174.219930203" lastFinishedPulling="2025-12-09 11:35:56.781765714 +0000 UTC m=+282.022524233" observedRunningTime="2025-12-09 11:36:08.345042297 +0000 UTC m=+293.585800816" watchObservedRunningTime="2025-12-09 11:36:26.16195052 +0000 UTC m=+311.402709039" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.163311 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cng6d" podStartSLOduration=34.913348285 podStartE2EDuration="2m24.163303775s" podCreationTimestamp="2025-12-09 11:34:02 +0000 UTC" firstStartedPulling="2025-12-09 11:34:07.608865318 +0000 UTC m=+172.849623837" lastFinishedPulling="2025-12-09 11:35:56.858820818 +0000 UTC m=+282.099579327" observedRunningTime="2025-12-09 11:36:08.365447575 +0000 UTC m=+293.606206084" watchObservedRunningTime="2025-12-09 11:36:26.163303775 +0000 UTC m=+311.404062294" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.164214 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.164208239 podStartE2EDuration="42.164208239s" podCreationTimestamp="2025-12-09 11:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:36:08.463695148 +0000 UTC m=+293.704453667" watchObservedRunningTime="2025-12-09 11:36:26.164208239 +0000 UTC m=+311.404966758" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.164398 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgrjv" podStartSLOduration=36.894996584 podStartE2EDuration="2m27.164393164s" podCreationTimestamp="2025-12-09 11:33:59 +0000 UTC" firstStartedPulling="2025-12-09 11:34:06.38822296 +0000 UTC m=+171.628981479" lastFinishedPulling="2025-12-09 11:35:56.65761954 +0000 UTC m=+281.898378059" observedRunningTime="2025-12-09 11:36:08.381207713 +0000 UTC m=+293.621966242" watchObservedRunningTime="2025-12-09 11:36:26.164393164 +0000 UTC m=+311.405151693" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.164498 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rpcqb" podStartSLOduration=36.292542199 podStartE2EDuration="2m24.164493786s" podCreationTimestamp="2025-12-09 11:34:02 +0000 UTC" firstStartedPulling="2025-12-09 11:34:08.986699267 +0000 UTC m=+174.227457786" lastFinishedPulling="2025-12-09 11:35:56.858650854 +0000 UTC m=+282.099409373" observedRunningTime="2025-12-09 11:36:08.267005757 +0000 UTC m=+293.507764286" watchObservedRunningTime="2025-12-09 11:36:26.164493786 +0000 UTC m=+311.405252305" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.165182 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-2wdxh"] Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.165243 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.169746 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.181700 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.217354 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.217334534 podStartE2EDuration="21.217334534s" podCreationTimestamp="2025-12-09 11:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:36:26.192476651 +0000 UTC m=+311.433235190" watchObservedRunningTime="2025-12-09 11:36:26.217334534 +0000 UTC m=+311.458093053" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.243086 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.280784 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.375304 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.422717 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.430682 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.480535 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.545726 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.695358 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.803222 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.806336 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.856717 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.880841 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.909180 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.927768 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.961893 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.961960 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.986710 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 11:36:26 crc kubenswrapper[4770]: I1209 11:36:26.993502 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.073770 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.083121 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.163221 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.172366 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.222328 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.327654 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.348608 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" path="/var/lib/kubelet/pods/e1a5a731-6de5-4cfa-abfd-b68487c0b855/volumes" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.358761 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.430568 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.441865 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.449883 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.480481 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.559601 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.600149 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.651150 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.720798 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.739262 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.747932 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.869190 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 11:36:27 crc kubenswrapper[4770]: I1209 11:36:27.944659 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.392647 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.421653 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.466236 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.523016 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.529703 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.626330 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.629669 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.654529 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.716613 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.742187 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.777810 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 11:36:28 crc kubenswrapper[4770]: I1209 11:36:28.790288 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.020337 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.024072 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.024147 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.038910 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.051286 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.105535 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.115218 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.126571 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.235349 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.394163 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.483834 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.488180 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.520500 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.580621 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.724434 4770 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.724789 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://cccd9cced28043456f1ddb11b6a254e0fd5e85cd65a4ddef5637118a773dd633" gracePeriod=5 Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.745625 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.828591 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.853427 4770 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 11:36:29 crc kubenswrapper[4770]: I1209 11:36:29.867973 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.088363 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.103429 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.118596 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.313934 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.405368 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.526679 4770 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.527163 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.557491 4770 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.607962 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.639039 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.656620 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.745840 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.749346 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.825387 4770 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.958739 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 11:36:30 crc kubenswrapper[4770]: I1209 11:36:30.997111 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.087339 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.329546 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.332836 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.340496 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.358691 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.376554 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.405657 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.459510 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.463459 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.470047 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.601855 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.688244 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.758746 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 11:36:31 crc kubenswrapper[4770]: I1209 11:36:31.983585 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.009076 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.080138 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.172720 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.184127 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.216153 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.227548 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.256989 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.575663 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.745684 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 11:36:32 crc kubenswrapper[4770]: I1209 11:36:32.956184 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.137303 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.169804 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.194329 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.241373 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248005 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-hg7tc"] Dec 09 11:36:33 crc kubenswrapper[4770]: E1209 11:36:33.248324 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248367 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 11:36:33 crc kubenswrapper[4770]: E1209 11:36:33.248396 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerName="oauth-openshift" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248406 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerName="oauth-openshift" Dec 09 11:36:33 crc kubenswrapper[4770]: E1209 11:36:33.248418 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" containerName="installer" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248427 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" containerName="installer" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248553 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1a5a731-6de5-4cfa-abfd-b68487c0b855" containerName="oauth-openshift" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248577 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e941ca15-c21b-4c0f-ba5b-3a8ea9ca99a7" containerName="installer" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.248589 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.249169 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256340 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256529 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256629 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256672 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256722 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.256660 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.257273 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.257407 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.257543 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.259042 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.259319 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.262407 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.285414 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.291111 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.300860 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.341277 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-dir\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.341382 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342027 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342543 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342597 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342631 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342733 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-policies\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342930 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv7xb\" (UniqueName: \"kubernetes.io/projected/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-kube-api-access-dv7xb\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.342985 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.343286 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.343449 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.343515 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.343586 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.343721 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444444 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444506 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444545 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-dir\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444667 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444697 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444748 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444767 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-policies\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv7xb\" (UniqueName: \"kubernetes.io/projected/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-kube-api-access-dv7xb\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.444865 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.445820 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-service-ca\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.445988 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-dir\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.446799 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-audit-policies\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.446821 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.448271 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.454639 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.454958 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-login\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.455777 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-router-certs\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.466101 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.470411 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-session\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.470411 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-error\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.471928 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.475471 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv7xb\" (UniqueName: \"kubernetes.io/projected/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-kube-api-access-dv7xb\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.478227 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/da7f7054-8d6f-48dc-8a31-1981a2b8f4e1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-684d58cf9b-hg7tc\" (UID: \"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1\") " pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.572505 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.621467 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.665802 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 11:36:33 crc kubenswrapper[4770]: I1209 11:36:33.758965 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.223763 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.237960 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.263645 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.426207 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.514196 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.619345 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.780788 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 11:36:34 crc kubenswrapper[4770]: I1209 11:36:34.960042 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.020646 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.163161 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.163238 4770 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="cccd9cced28043456f1ddb11b6a254e0fd5e85cd65a4ddef5637118a773dd633" exitCode=137 Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.315884 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.316027 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.347702 4770 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.361355 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.361449 4770 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b60f9f08-2b01-409f-87c3-3f35347ba8ee" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.369021 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.369063 4770 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b60f9f08-2b01-409f-87c3-3f35347ba8ee" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.474594 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.474691 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.474735 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.474787 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.474889 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.475537 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.475625 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.476341 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.476376 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.488187 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.577116 4770 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.577165 4770 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.577178 4770 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.577190 4770 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:35 crc kubenswrapper[4770]: I1209 11:36:35.577202 4770 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:36 crc kubenswrapper[4770]: I1209 11:36:36.170612 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 11:36:36 crc kubenswrapper[4770]: I1209 11:36:36.170704 4770 scope.go:117] "RemoveContainer" containerID="cccd9cced28043456f1ddb11b6a254e0fd5e85cd65a4ddef5637118a773dd633" Dec 09 11:36:36 crc kubenswrapper[4770]: I1209 11:36:36.170814 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 11:36:37 crc kubenswrapper[4770]: I1209 11:36:37.347433 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 09 11:36:49 crc kubenswrapper[4770]: I1209 11:36:49.308418 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.075934 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.076423 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" podUID="bd6aba04-bce5-4231-b2c5-c3b574409d90" containerName="controller-manager" containerID="cri-o://4bfd8733cc57675215c929e31b7cd61e4c7d8187f617d8dc5ea1142f71a0b475" gracePeriod=30 Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.259868 4770 generic.go:334] "Generic (PLEG): container finished" podID="bd6aba04-bce5-4231-b2c5-c3b574409d90" containerID="4bfd8733cc57675215c929e31b7cd61e4c7d8187f617d8dc5ea1142f71a0b475" exitCode=0 Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.259940 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" event={"ID":"bd6aba04-bce5-4231-b2c5-c3b574409d90","Type":"ContainerDied","Data":"4bfd8733cc57675215c929e31b7cd61e4c7d8187f617d8dc5ea1142f71a0b475"} Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.277507 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.277820 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" containerID="cri-o://5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9" gracePeriod=30 Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.542784 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.614878 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.706713 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert\") pod \"bd6aba04-bce5-4231-b2c5-c3b574409d90\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.706781 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca\") pod \"bd6aba04-bce5-4231-b2c5-c3b574409d90\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.706870 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config\") pod \"bd6aba04-bce5-4231-b2c5-c3b574409d90\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.706976 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g27k9\" (UniqueName: \"kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9\") pod \"bd6aba04-bce5-4231-b2c5-c3b574409d90\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.707088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles\") pod \"bd6aba04-bce5-4231-b2c5-c3b574409d90\" (UID: \"bd6aba04-bce5-4231-b2c5-c3b574409d90\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.708613 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca" (OuterVolumeSpecName: "client-ca") pod "bd6aba04-bce5-4231-b2c5-c3b574409d90" (UID: "bd6aba04-bce5-4231-b2c5-c3b574409d90"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.708650 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bd6aba04-bce5-4231-b2c5-c3b574409d90" (UID: "bd6aba04-bce5-4231-b2c5-c3b574409d90"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.708680 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config" (OuterVolumeSpecName: "config") pod "bd6aba04-bce5-4231-b2c5-c3b574409d90" (UID: "bd6aba04-bce5-4231-b2c5-c3b574409d90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.714488 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bd6aba04-bce5-4231-b2c5-c3b574409d90" (UID: "bd6aba04-bce5-4231-b2c5-c3b574409d90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.714578 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9" (OuterVolumeSpecName: "kube-api-access-g27k9") pod "bd6aba04-bce5-4231-b2c5-c3b574409d90" (UID: "bd6aba04-bce5-4231-b2c5-c3b574409d90"). InnerVolumeSpecName "kube-api-access-g27k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert\") pod \"ff350dff-877d-4fac-8ed7-322c0e8894e0\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808380 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca\") pod \"ff350dff-877d-4fac-8ed7-322c0e8894e0\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808489 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dj9n\" (UniqueName: \"kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n\") pod \"ff350dff-877d-4fac-8ed7-322c0e8894e0\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808534 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config\") pod \"ff350dff-877d-4fac-8ed7-322c0e8894e0\" (UID: \"ff350dff-877d-4fac-8ed7-322c0e8894e0\") " Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808889 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd6aba04-bce5-4231-b2c5-c3b574409d90-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808934 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808946 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808959 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g27k9\" (UniqueName: \"kubernetes.io/projected/bd6aba04-bce5-4231-b2c5-c3b574409d90-kube-api-access-g27k9\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.808973 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bd6aba04-bce5-4231-b2c5-c3b574409d90-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.809690 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config" (OuterVolumeSpecName: "config") pod "ff350dff-877d-4fac-8ed7-322c0e8894e0" (UID: "ff350dff-877d-4fac-8ed7-322c0e8894e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.809675 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca" (OuterVolumeSpecName: "client-ca") pod "ff350dff-877d-4fac-8ed7-322c0e8894e0" (UID: "ff350dff-877d-4fac-8ed7-322c0e8894e0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.813462 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ff350dff-877d-4fac-8ed7-322c0e8894e0" (UID: "ff350dff-877d-4fac-8ed7-322c0e8894e0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.814677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n" (OuterVolumeSpecName: "kube-api-access-8dj9n") pod "ff350dff-877d-4fac-8ed7-322c0e8894e0" (UID: "ff350dff-877d-4fac-8ed7-322c0e8894e0"). InnerVolumeSpecName "kube-api-access-8dj9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.898160 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.910787 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff350dff-877d-4fac-8ed7-322c0e8894e0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.910857 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.910868 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dj9n\" (UniqueName: \"kubernetes.io/projected/ff350dff-877d-4fac-8ed7-322c0e8894e0-kube-api-access-8dj9n\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:50 crc kubenswrapper[4770]: I1209 11:36:50.910885 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff350dff-877d-4fac-8ed7-322c0e8894e0-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.270330 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" event={"ID":"bd6aba04-bce5-4231-b2c5-c3b574409d90","Type":"ContainerDied","Data":"8295b130c99283bb322cfb99fcc1fd36370d1449e1064592202c9b9ef35571c7"} Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.270402 4770 scope.go:117] "RemoveContainer" containerID="4bfd8733cc57675215c929e31b7cd61e4c7d8187f617d8dc5ea1142f71a0b475" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.270350 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-frwww" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.272823 4770 generic.go:334] "Generic (PLEG): container finished" podID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerID="5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9" exitCode=0 Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.272871 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" event={"ID":"ff350dff-877d-4fac-8ed7-322c0e8894e0","Type":"ContainerDied","Data":"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9"} Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.272895 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.272929 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw" event={"ID":"ff350dff-877d-4fac-8ed7-322c0e8894e0","Type":"ContainerDied","Data":"380918ecfd2f605c8a9a3f4cf11838b014e6d05d715ff43cffc62a3b62fa037a"} Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.291820 4770 scope.go:117] "RemoveContainer" containerID="5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.308490 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.316235 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-frwww"] Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.316353 4770 scope.go:117] "RemoveContainer" containerID="5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9" Dec 09 11:36:51 crc kubenswrapper[4770]: E1209 11:36:51.317484 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9\": container with ID starting with 5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9 not found: ID does not exist" containerID="5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.317552 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9"} err="failed to get container status \"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9\": rpc error: code = NotFound desc = could not find container \"5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9\": container with ID starting with 5094b91739c5948b52186f1ebdcb967bf793c8728d91fe87c70d1ba1da38c2c9 not found: ID does not exist" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.321811 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.326939 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8ftw"] Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.352467 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd6aba04-bce5-4231-b2c5-c3b574409d90" path="/var/lib/kubelet/pods/bd6aba04-bce5-4231-b2c5-c3b574409d90/volumes" Dec 09 11:36:51 crc kubenswrapper[4770]: I1209 11:36:51.353159 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" path="/var/lib/kubelet/pods/ff350dff-877d-4fac-8ed7-322c0e8894e0/volumes" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.259353 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:36:52 crc kubenswrapper[4770]: E1209 11:36:52.260145 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6aba04-bce5-4231-b2c5-c3b574409d90" containerName="controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.260165 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6aba04-bce5-4231-b2c5-c3b574409d90" containerName="controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: E1209 11:36:52.260191 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.260197 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.260327 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6aba04-bce5-4231-b2c5-c3b574409d90" containerName="controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.260349 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff350dff-877d-4fac-8ed7-322c0e8894e0" containerName="route-controller-manager" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.261020 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.263866 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.264080 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b"] Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.265186 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.268302 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.268633 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.268980 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.268991 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269186 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269389 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269598 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269740 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269886 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.269987 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.270847 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.273570 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.332541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.332597 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.332650 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmt6\" (UniqueName: \"kubernetes.io/projected/9a86d378-4e7a-408d-8c9a-bc878235a380-kube-api-access-gcmt6\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.332674 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a86d378-4e7a-408d-8c9a-bc878235a380-serving-cert\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.332692 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.334229 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpx4m\" (UniqueName: \"kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.334275 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-client-ca\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.334325 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.334375 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-config\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435386 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmt6\" (UniqueName: \"kubernetes.io/projected/9a86d378-4e7a-408d-8c9a-bc878235a380-kube-api-access-gcmt6\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435454 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a86d378-4e7a-408d-8c9a-bc878235a380-serving-cert\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435490 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpx4m\" (UniqueName: \"kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435543 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-client-ca\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435586 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435617 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-config\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.435664 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.437768 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.438115 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-client-ca\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.438242 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.438265 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.439015 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a86d378-4e7a-408d-8c9a-bc878235a380-config\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.439560 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.444135 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a86d378-4e7a-408d-8c9a-bc878235a380-serving-cert\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.444153 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.460985 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpx4m\" (UniqueName: \"kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m\") pod \"controller-manager-7cc945dc9b-mfxnk\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.461338 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmt6\" (UniqueName: \"kubernetes.io/projected/9a86d378-4e7a-408d-8c9a-bc878235a380-kube-api-access-gcmt6\") pod \"route-controller-manager-6b694b59cf-qwc7b\" (UID: \"9a86d378-4e7a-408d-8c9a-bc878235a380\") " pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.583166 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:36:52 crc kubenswrapper[4770]: I1209 11:36:52.594627 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:36:53 crc kubenswrapper[4770]: I1209 11:36:53.947072 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 11:36:54 crc kubenswrapper[4770]: I1209 11:36:54.708127 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 11:36:56 crc kubenswrapper[4770]: I1209 11:36:56.015300 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 11:36:57 crc kubenswrapper[4770]: I1209 11:36:57.557562 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 11:36:58 crc kubenswrapper[4770]: I1209 11:36:58.006280 4770 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 11:36:58 crc kubenswrapper[4770]: I1209 11:36:58.956385 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 11:37:00 crc kubenswrapper[4770]: I1209 11:37:00.938496 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 11:37:02 crc kubenswrapper[4770]: I1209 11:37:02.599785 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.167154 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.450987 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.462839 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-hg7tc"] Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.480598 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b"] Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.641315 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b"] Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.826518 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-684d58cf9b-hg7tc"] Dec 09 11:37:09 crc kubenswrapper[4770]: I1209 11:37:09.882463 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:37:09 crc kubenswrapper[4770]: W1209 11:37:09.891429 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aa70b40_980f_4bac_98b0_338160c31bc9.slice/crio-7ab5b0f3d1c286f1daf61d5c846e33eb1bc5eff275c7c5bd726989746f549666 WatchSource:0}: Error finding container 7ab5b0f3d1c286f1daf61d5c846e33eb1bc5eff275c7c5bd726989746f549666: Status 404 returned error can't find the container with id 7ab5b0f3d1c286f1daf61d5c846e33eb1bc5eff275c7c5bd726989746f549666 Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.041264 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.464356 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" event={"ID":"9a86d378-4e7a-408d-8c9a-bc878235a380","Type":"ContainerStarted","Data":"6a139aea5bec4c185a7bb10dbdfeba4a3757035e0f33a5a40526ec8f2e7f40c3"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.464609 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" event={"ID":"9a86d378-4e7a-408d-8c9a-bc878235a380","Type":"ContainerStarted","Data":"3f7f46c76020e04f22f30b701f59ee6dd2b59d45acfaf4efaf94b1fedda0ccff"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.467744 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.470704 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" event={"ID":"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1","Type":"ContainerStarted","Data":"f44e25e6da3121a64227a7823b40570a4e5a168d0a5a1fd8ad703ca9b544252c"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.470739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" event={"ID":"da7f7054-8d6f-48dc-8a31-1981a2b8f4e1","Type":"ContainerStarted","Data":"76412d6673293535446aac1be1af5238a13928ebd3b77987fd85e5473708321e"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.473198 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.479161 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" event={"ID":"2aa70b40-980f-4bac-98b0-338160c31bc9","Type":"ContainerStarted","Data":"1eef4685d9617a0f8eadc0050350a779c214db23e4b0b983edb019752efff099"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.479546 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" event={"ID":"2aa70b40-980f-4bac-98b0-338160c31bc9","Type":"ContainerStarted","Data":"7ab5b0f3d1c286f1daf61d5c846e33eb1bc5eff275c7c5bd726989746f549666"} Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.479855 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" podUID="2aa70b40-980f-4bac-98b0-338160c31bc9" containerName="controller-manager" containerID="cri-o://1eef4685d9617a0f8eadc0050350a779c214db23e4b0b983edb019752efff099" gracePeriod=30 Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.482662 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.483409 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.506335 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.514067 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.526785 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b694b59cf-qwc7b" podStartSLOduration=20.52675003 podStartE2EDuration="20.52675003s" podCreationTimestamp="2025-12-09 11:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:37:10.49758577 +0000 UTC m=+355.738344289" watchObservedRunningTime="2025-12-09 11:37:10.52675003 +0000 UTC m=+355.767508549" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.528551 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" podStartSLOduration=20.528537027 podStartE2EDuration="20.528537027s" podCreationTimestamp="2025-12-09 11:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:37:10.524357458 +0000 UTC m=+355.765115987" watchObservedRunningTime="2025-12-09 11:37:10.528537027 +0000 UTC m=+355.769295546" Dec 09 11:37:10 crc kubenswrapper[4770]: I1209 11:37:10.547205 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-684d58cf9b-hg7tc" podStartSLOduration=88.547189873 podStartE2EDuration="1m28.547189873s" podCreationTimestamp="2025-12-09 11:35:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:37:10.546531596 +0000 UTC m=+355.787290115" watchObservedRunningTime="2025-12-09 11:37:10.547189873 +0000 UTC m=+355.787948392" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.491807 4770 generic.go:334] "Generic (PLEG): container finished" podID="2aa70b40-980f-4bac-98b0-338160c31bc9" containerID="1eef4685d9617a0f8eadc0050350a779c214db23e4b0b983edb019752efff099" exitCode=0 Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.493155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" event={"ID":"2aa70b40-980f-4bac-98b0-338160c31bc9","Type":"ContainerDied","Data":"1eef4685d9617a0f8eadc0050350a779c214db23e4b0b983edb019752efff099"} Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.908015 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.941499 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk"] Dec 09 11:37:11 crc kubenswrapper[4770]: E1209 11:37:11.941851 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa70b40-980f-4bac-98b0-338160c31bc9" containerName="controller-manager" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.941868 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa70b40-980f-4bac-98b0-338160c31bc9" containerName="controller-manager" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.942020 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa70b40-980f-4bac-98b0-338160c31bc9" containerName="controller-manager" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.942530 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.969009 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk"] Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983328 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca\") pod \"2aa70b40-980f-4bac-98b0-338160c31bc9\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983428 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles\") pod \"2aa70b40-980f-4bac-98b0-338160c31bc9\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983492 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config\") pod \"2aa70b40-980f-4bac-98b0-338160c31bc9\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983604 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert\") pod \"2aa70b40-980f-4bac-98b0-338160c31bc9\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpx4m\" (UniqueName: \"kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m\") pod \"2aa70b40-980f-4bac-98b0-338160c31bc9\" (UID: \"2aa70b40-980f-4bac-98b0-338160c31bc9\") " Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983923 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4mn8\" (UniqueName: \"kubernetes.io/projected/9204283f-a642-40f8-ac41-8f2b49aafcad-kube-api-access-h4mn8\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.983996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-client-ca\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.984025 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-proxy-ca-bundles\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.984074 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9204283f-a642-40f8-ac41-8f2b49aafcad-serving-cert\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.984110 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-config\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.985590 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "2aa70b40-980f-4bac-98b0-338160c31bc9" (UID: "2aa70b40-980f-4bac-98b0-338160c31bc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.985705 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config" (OuterVolumeSpecName: "config") pod "2aa70b40-980f-4bac-98b0-338160c31bc9" (UID: "2aa70b40-980f-4bac-98b0-338160c31bc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.985627 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2aa70b40-980f-4bac-98b0-338160c31bc9" (UID: "2aa70b40-980f-4bac-98b0-338160c31bc9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.991834 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2aa70b40-980f-4bac-98b0-338160c31bc9" (UID: "2aa70b40-980f-4bac-98b0-338160c31bc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:37:11 crc kubenswrapper[4770]: I1209 11:37:11.999142 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m" (OuterVolumeSpecName: "kube-api-access-zpx4m") pod "2aa70b40-980f-4bac-98b0-338160c31bc9" (UID: "2aa70b40-980f-4bac-98b0-338160c31bc9"). InnerVolumeSpecName "kube-api-access-zpx4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086063 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9204283f-a642-40f8-ac41-8f2b49aafcad-serving-cert\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086169 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-config\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4mn8\" (UniqueName: \"kubernetes.io/projected/9204283f-a642-40f8-ac41-8f2b49aafcad-kube-api-access-h4mn8\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-client-ca\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086310 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-proxy-ca-bundles\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086373 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086389 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086402 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa70b40-980f-4bac-98b0-338160c31bc9-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086414 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2aa70b40-980f-4bac-98b0-338160c31bc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.086426 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpx4m\" (UniqueName: \"kubernetes.io/projected/2aa70b40-980f-4bac-98b0-338160c31bc9-kube-api-access-zpx4m\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.088483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-config\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.088556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-client-ca\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.088612 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9204283f-a642-40f8-ac41-8f2b49aafcad-proxy-ca-bundles\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.092177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9204283f-a642-40f8-ac41-8f2b49aafcad-serving-cert\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.110422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4mn8\" (UniqueName: \"kubernetes.io/projected/9204283f-a642-40f8-ac41-8f2b49aafcad-kube-api-access-h4mn8\") pod \"controller-manager-75c8cbdcf5-xbznk\" (UID: \"9204283f-a642-40f8-ac41-8f2b49aafcad\") " pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.267369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.525101 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" event={"ID":"2aa70b40-980f-4bac-98b0-338160c31bc9","Type":"ContainerDied","Data":"7ab5b0f3d1c286f1daf61d5c846e33eb1bc5eff275c7c5bd726989746f549666"} Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.525780 4770 scope.go:117] "RemoveContainer" containerID="1eef4685d9617a0f8eadc0050350a779c214db23e4b0b983edb019752efff099" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.525155 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk" Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.584234 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.588391 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cc945dc9b-mfxnk"] Dec 09 11:37:12 crc kubenswrapper[4770]: I1209 11:37:12.599133 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk"] Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.351078 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa70b40-980f-4bac-98b0-338160c31bc9" path="/var/lib/kubelet/pods/2aa70b40-980f-4bac-98b0-338160c31bc9/volumes" Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.353676 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.354035 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ktnmb" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="registry-server" containerID="cri-o://d5525354f02add30b14fcf41c39432ca518eeaca5218eff280dbfed5e8708e02" gracePeriod=2 Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.537670 4770 generic.go:334] "Generic (PLEG): container finished" podID="92fca08c-4561-4733-a9d2-488e013b0da1" containerID="d5525354f02add30b14fcf41c39432ca518eeaca5218eff280dbfed5e8708e02" exitCode=0 Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.537791 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerDied","Data":"d5525354f02add30b14fcf41c39432ca518eeaca5218eff280dbfed5e8708e02"} Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.553521 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.554730 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6z5lm" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="registry-server" containerID="cri-o://07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de" gracePeriod=2 Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.560951 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" event={"ID":"9204283f-a642-40f8-ac41-8f2b49aafcad","Type":"ContainerStarted","Data":"70f6d0148dc11580e4bb83820555e8ffa4d3d2cb7f3ac943120c930d9a758518"} Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.561029 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" event={"ID":"9204283f-a642-40f8-ac41-8f2b49aafcad","Type":"ContainerStarted","Data":"6b3be98d96eeb791574ac7a8c051f70fef86dbcfe5b437a19b6280fc52dbe8ff"} Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.561511 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.566788 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.581993 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75c8cbdcf5-xbznk" podStartSLOduration=3.581972873 podStartE2EDuration="3.581972873s" podCreationTimestamp="2025-12-09 11:37:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:37:13.580419823 +0000 UTC m=+358.821178362" watchObservedRunningTime="2025-12-09 11:37:13.581972873 +0000 UTC m=+358.822731392" Dec 09 11:37:13 crc kubenswrapper[4770]: I1209 11:37:13.825192 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.010708 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities\") pod \"92fca08c-4561-4733-a9d2-488e013b0da1\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.011334 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgqmq\" (UniqueName: \"kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq\") pod \"92fca08c-4561-4733-a9d2-488e013b0da1\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.011427 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content\") pod \"92fca08c-4561-4733-a9d2-488e013b0da1\" (UID: \"92fca08c-4561-4733-a9d2-488e013b0da1\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.012102 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities" (OuterVolumeSpecName: "utilities") pod "92fca08c-4561-4733-a9d2-488e013b0da1" (UID: "92fca08c-4561-4733-a9d2-488e013b0da1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.023060 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq" (OuterVolumeSpecName: "kube-api-access-dgqmq") pod "92fca08c-4561-4733-a9d2-488e013b0da1" (UID: "92fca08c-4561-4733-a9d2-488e013b0da1"). InnerVolumeSpecName "kube-api-access-dgqmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.061663 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.080759 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92fca08c-4561-4733-a9d2-488e013b0da1" (UID: "92fca08c-4561-4733-a9d2-488e013b0da1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112147 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhpsf\" (UniqueName: \"kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf\") pod \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112215 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content\") pod \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112237 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities\") pod \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\" (UID: \"7ffafda4-fd6e-4831-a20e-f50e5232a3ba\") " Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112408 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112421 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgqmq\" (UniqueName: \"kubernetes.io/projected/92fca08c-4561-4733-a9d2-488e013b0da1-kube-api-access-dgqmq\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.112432 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fca08c-4561-4733-a9d2-488e013b0da1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.113224 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities" (OuterVolumeSpecName: "utilities") pod "7ffafda4-fd6e-4831-a20e-f50e5232a3ba" (UID: "7ffafda4-fd6e-4831-a20e-f50e5232a3ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.125346 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf" (OuterVolumeSpecName: "kube-api-access-qhpsf") pod "7ffafda4-fd6e-4831-a20e-f50e5232a3ba" (UID: "7ffafda4-fd6e-4831-a20e-f50e5232a3ba"). InnerVolumeSpecName "kube-api-access-qhpsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.167661 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ffafda4-fd6e-4831-a20e-f50e5232a3ba" (UID: "7ffafda4-fd6e-4831-a20e-f50e5232a3ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.214043 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhpsf\" (UniqueName: \"kubernetes.io/projected/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-kube-api-access-qhpsf\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.214118 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.214136 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ffafda4-fd6e-4831-a20e-f50e5232a3ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.570190 4770 generic.go:334] "Generic (PLEG): container finished" podID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerID="07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de" exitCode=0 Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.570277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerDied","Data":"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de"} Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.570319 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5lm" event={"ID":"7ffafda4-fd6e-4831-a20e-f50e5232a3ba","Type":"ContainerDied","Data":"ce7966742fe50f6d6375369f29e47625148027df3374b776249200da9001e8f8"} Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.570311 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5lm" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.570379 4770 scope.go:117] "RemoveContainer" containerID="07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.579052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktnmb" event={"ID":"92fca08c-4561-4733-a9d2-488e013b0da1","Type":"ContainerDied","Data":"abd162848a0002a3568f5681c29b73cfc2abf2bd93853f3e6326ef2d3f32dc8d"} Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.579083 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktnmb" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.589039 4770 scope.go:117] "RemoveContainer" containerID="66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.614025 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.618430 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6z5lm"] Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.638853 4770 scope.go:117] "RemoveContainer" containerID="8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.653446 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.656823 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ktnmb"] Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.663500 4770 scope.go:117] "RemoveContainer" containerID="07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de" Dec 09 11:37:14 crc kubenswrapper[4770]: E1209 11:37:14.664094 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de\": container with ID starting with 07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de not found: ID does not exist" containerID="07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.664144 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de"} err="failed to get container status \"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de\": rpc error: code = NotFound desc = could not find container \"07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de\": container with ID starting with 07f63b458a5f27f463762a081006ea70c4e61e608b9bb8f214f75f47de2c96de not found: ID does not exist" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.664186 4770 scope.go:117] "RemoveContainer" containerID="66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9" Dec 09 11:37:14 crc kubenswrapper[4770]: E1209 11:37:14.664628 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9\": container with ID starting with 66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9 not found: ID does not exist" containerID="66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.664800 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9"} err="failed to get container status \"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9\": rpc error: code = NotFound desc = could not find container \"66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9\": container with ID starting with 66a368922f399b2d566bc5eb5251e1e553c902ea1f0b709177841ec83ad61fb9 not found: ID does not exist" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.664958 4770 scope.go:117] "RemoveContainer" containerID="8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67" Dec 09 11:37:14 crc kubenswrapper[4770]: E1209 11:37:14.665834 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67\": container with ID starting with 8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67 not found: ID does not exist" containerID="8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.665862 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67"} err="failed to get container status \"8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67\": rpc error: code = NotFound desc = could not find container \"8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67\": container with ID starting with 8de37e81b9bf399ab76f2799ad3eeea2a0c8bb437cb0c6dc71661c29eeaf9b67 not found: ID does not exist" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.665880 4770 scope.go:117] "RemoveContainer" containerID="d5525354f02add30b14fcf41c39432ca518eeaca5218eff280dbfed5e8708e02" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.687021 4770 scope.go:117] "RemoveContainer" containerID="e9ee728d3985ba497c6667c0eb76c6d97f150138f447a02a8d43112874eac6db" Dec 09 11:37:14 crc kubenswrapper[4770]: I1209 11:37:14.714616 4770 scope.go:117] "RemoveContainer" containerID="1be4e0aa0fa7ea65d247e0e2b46f3bec0abe9155d60b1b5e84138cd34b3dcb47" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.151855 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.152472 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fg6m5" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="registry-server" containerID="cri-o://cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38" gracePeriod=2 Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.348191 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" path="/var/lib/kubelet/pods/7ffafda4-fd6e-4831-a20e-f50e5232a3ba/volumes" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.349464 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" path="/var/lib/kubelet/pods/92fca08c-4561-4733-a9d2-488e013b0da1/volumes" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.535087 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.588440 4770 generic.go:334] "Generic (PLEG): container finished" podID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerID="cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38" exitCode=0 Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.588548 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg6m5" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.588603 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerDied","Data":"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38"} Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.588719 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg6m5" event={"ID":"0a151ffe-1d66-4d45-bd04-604d30df30a1","Type":"ContainerDied","Data":"d31bdaf35e7b12b5dffc8044d22bc441d382b2abf701c8805455aa9256d7210d"} Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.588752 4770 scope.go:117] "RemoveContainer" containerID="cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.605605 4770 scope.go:117] "RemoveContainer" containerID="09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.623393 4770 scope.go:117] "RemoveContainer" containerID="fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.638620 4770 scope.go:117] "RemoveContainer" containerID="cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.639062 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content\") pod \"0a151ffe-1d66-4d45-bd04-604d30df30a1\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.639224 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b7cq\" (UniqueName: \"kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq\") pod \"0a151ffe-1d66-4d45-bd04-604d30df30a1\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.639267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities\") pod \"0a151ffe-1d66-4d45-bd04-604d30df30a1\" (UID: \"0a151ffe-1d66-4d45-bd04-604d30df30a1\") " Dec 09 11:37:15 crc kubenswrapper[4770]: E1209 11:37:15.639331 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38\": container with ID starting with cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38 not found: ID does not exist" containerID="cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.639383 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38"} err="failed to get container status \"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38\": rpc error: code = NotFound desc = could not find container \"cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38\": container with ID starting with cac59e48eace84845472a249b0448c5d58259f26843f15e65043e2b3ad79ee38 not found: ID does not exist" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.639425 4770 scope.go:117] "RemoveContainer" containerID="09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.640568 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities" (OuterVolumeSpecName: "utilities") pod "0a151ffe-1d66-4d45-bd04-604d30df30a1" (UID: "0a151ffe-1d66-4d45-bd04-604d30df30a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:15 crc kubenswrapper[4770]: E1209 11:37:15.640836 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991\": container with ID starting with 09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991 not found: ID does not exist" containerID="09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.640880 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991"} err="failed to get container status \"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991\": rpc error: code = NotFound desc = could not find container \"09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991\": container with ID starting with 09c05ed25a629a56f7f7e0283b20fd618b606c552e668a9f508abdf30e0b5991 not found: ID does not exist" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.640938 4770 scope.go:117] "RemoveContainer" containerID="fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075" Dec 09 11:37:15 crc kubenswrapper[4770]: E1209 11:37:15.641261 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075\": container with ID starting with fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075 not found: ID does not exist" containerID="fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.641287 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075"} err="failed to get container status \"fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075\": rpc error: code = NotFound desc = could not find container \"fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075\": container with ID starting with fe5e8a16d0c1e1ab74f78eacc2d47254fca838aa2721851be05b3b4f532b2075 not found: ID does not exist" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.645639 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq" (OuterVolumeSpecName: "kube-api-access-5b7cq") pod "0a151ffe-1d66-4d45-bd04-604d30df30a1" (UID: "0a151ffe-1d66-4d45-bd04-604d30df30a1"). InnerVolumeSpecName "kube-api-access-5b7cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.664083 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a151ffe-1d66-4d45-bd04-604d30df30a1" (UID: "0a151ffe-1d66-4d45-bd04-604d30df30a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.741135 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b7cq\" (UniqueName: \"kubernetes.io/projected/0a151ffe-1d66-4d45-bd04-604d30df30a1-kube-api-access-5b7cq\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.741178 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.741195 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a151ffe-1d66-4d45-bd04-604d30df30a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.929356 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:37:15 crc kubenswrapper[4770]: I1209 11:37:15.934543 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg6m5"] Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.353867 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.354157 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rpcqb" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="registry-server" containerID="cri-o://e94ffe2a3195b5b5a16e2cb8a93faaac3f88681c7e9d1185493e88fb1c78e1d1" gracePeriod=2 Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.618593 4770 generic.go:334] "Generic (PLEG): container finished" podID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerID="e94ffe2a3195b5b5a16e2cb8a93faaac3f88681c7e9d1185493e88fb1c78e1d1" exitCode=0 Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.618662 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerDied","Data":"e94ffe2a3195b5b5a16e2cb8a93faaac3f88681c7e9d1185493e88fb1c78e1d1"} Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.743800 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.854728 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4vc6\" (UniqueName: \"kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6\") pod \"357916a9-0349-4b3c-b3c2-58af2ee26540\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.854797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content\") pod \"357916a9-0349-4b3c-b3c2-58af2ee26540\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.854820 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities\") pod \"357916a9-0349-4b3c-b3c2-58af2ee26540\" (UID: \"357916a9-0349-4b3c-b3c2-58af2ee26540\") " Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.855829 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities" (OuterVolumeSpecName: "utilities") pod "357916a9-0349-4b3c-b3c2-58af2ee26540" (UID: "357916a9-0349-4b3c-b3c2-58af2ee26540"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.859563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6" (OuterVolumeSpecName: "kube-api-access-h4vc6") pod "357916a9-0349-4b3c-b3c2-58af2ee26540" (UID: "357916a9-0349-4b3c-b3c2-58af2ee26540"). InnerVolumeSpecName "kube-api-access-h4vc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.956803 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4vc6\" (UniqueName: \"kubernetes.io/projected/357916a9-0349-4b3c-b3c2-58af2ee26540-kube-api-access-h4vc6\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.956865 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:16 crc kubenswrapper[4770]: I1209 11:37:16.972269 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "357916a9-0349-4b3c-b3c2-58af2ee26540" (UID: "357916a9-0349-4b3c-b3c2-58af2ee26540"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.057775 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/357916a9-0349-4b3c-b3c2-58af2ee26540-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.348409 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" path="/var/lib/kubelet/pods/0a151ffe-1d66-4d45-bd04-604d30df30a1/volumes" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.627485 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpcqb" event={"ID":"357916a9-0349-4b3c-b3c2-58af2ee26540","Type":"ContainerDied","Data":"88ffc329ced6accd5d6af6939b9a086e012da28293b001cc5f112071111a7e76"} Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.627542 4770 scope.go:117] "RemoveContainer" containerID="e94ffe2a3195b5b5a16e2cb8a93faaac3f88681c7e9d1185493e88fb1c78e1d1" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.627658 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpcqb" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.647131 4770 scope.go:117] "RemoveContainer" containerID="639175d28e23ccd2f22f5209a68cb92e6624f9dbfebe1e590ad2d9432db5edef" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.661992 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.670315 4770 scope.go:117] "RemoveContainer" containerID="ec18afe402bdd1f5f778fb6773869b4ad784e8d71a29f441556bb68f7f927bb8" Dec 09 11:37:17 crc kubenswrapper[4770]: I1209 11:37:17.684611 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rpcqb"] Dec 09 11:37:19 crc kubenswrapper[4770]: I1209 11:37:19.349129 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" path="/var/lib/kubelet/pods/357916a9-0349-4b3c-b3c2-58af2ee26540/volumes" Dec 09 11:37:32 crc kubenswrapper[4770]: I1209 11:37:32.483194 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:37:32 crc kubenswrapper[4770]: I1209 11:37:32.483866 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.402828 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.404301 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-84dm9" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="registry-server" containerID="cri-o://9186eca8ab5b6b6f088b79a4bedaf5039e7602d97c2cfeb8e736335bf0ff68f9" gracePeriod=30 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.419273 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.419631 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgrjv" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="registry-server" containerID="cri-o://7a0bfea976d7b11910250169c1146afa4ec8aaaa7a9d9baf008fad3ceb4c65a7" gracePeriod=30 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.431032 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.431323 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" containerID="cri-o://7da83339064de66aef2ffeeb28a472af93299d9111a3a0d6ed18c271c2cce8b4" gracePeriod=30 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.439933 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.440261 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vnc6l" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="registry-server" containerID="cri-o://32b37d1cd3dc61b78afc9fd2df735d736e635d9faa7feabada5417cb9180034c" gracePeriod=30 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.449259 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.449539 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cng6d" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="registry-server" containerID="cri-o://028a69e195d308587c22bd9b31ad909606b0269f7c88819cd17e0a2731e821d3" gracePeriod=30 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.472790 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-twm9p"] Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473155 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473172 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473182 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473193 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473202 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473221 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473237 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473242 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473250 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473256 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473265 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473271 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473284 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473290 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473300 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473305 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473316 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473322 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473335 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473341 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="extract-utilities" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473348 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473354 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: E1209 11:37:54.473364 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473369 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="extract-content" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473480 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a151ffe-1d66-4d45-bd04-604d30df30a1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473496 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="92fca08c-4561-4733-a9d2-488e013b0da1" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473504 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ffafda4-fd6e-4831-a20e-f50e5232a3ba" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.473510 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="357916a9-0349-4b3c-b3c2-58af2ee26540" containerName="registry-server" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.474050 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.495703 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-twm9p"] Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.496618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.496658 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.496700 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtn8z\" (UniqueName: \"kubernetes.io/projected/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-kube-api-access-xtn8z\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.597586 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.597654 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.597701 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtn8z\" (UniqueName: \"kubernetes.io/projected/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-kube-api-access-xtn8z\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.600444 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.611825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.614709 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtn8z\" (UniqueName: \"kubernetes.io/projected/9e029e5b-b9b0-483a-ad49-29f8ceb43e61-kube-api-access-xtn8z\") pod \"marketplace-operator-79b997595-twm9p\" (UID: \"9e029e5b-b9b0-483a-ad49-29f8ceb43e61\") " pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.870222 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerID="9186eca8ab5b6b6f088b79a4bedaf5039e7602d97c2cfeb8e736335bf0ff68f9" exitCode=0 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.870474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerDied","Data":"9186eca8ab5b6b6f088b79a4bedaf5039e7602d97c2cfeb8e736335bf0ff68f9"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.870593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84dm9" event={"ID":"b9afd8a5-f75a-4aa5-8cba-a4699b25b847","Type":"ContainerDied","Data":"ff7778012e3751cd84152cbf29e5f28b702b885c2d5123d417e4272121bd208d"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.870613 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff7778012e3751cd84152cbf29e5f28b702b885c2d5123d417e4272121bd208d" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.874043 4770 generic.go:334] "Generic (PLEG): container finished" podID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerID="7a0bfea976d7b11910250169c1146afa4ec8aaaa7a9d9baf008fad3ceb4c65a7" exitCode=0 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.874107 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerDied","Data":"7a0bfea976d7b11910250169c1146afa4ec8aaaa7a9d9baf008fad3ceb4c65a7"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.877845 4770 generic.go:334] "Generic (PLEG): container finished" podID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerID="32b37d1cd3dc61b78afc9fd2df735d736e635d9faa7feabada5417cb9180034c" exitCode=0 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.877920 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerDied","Data":"32b37d1cd3dc61b78afc9fd2df735d736e635d9faa7feabada5417cb9180034c"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.887477 4770 generic.go:334] "Generic (PLEG): container finished" podID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerID="7da83339064de66aef2ffeeb28a472af93299d9111a3a0d6ed18c271c2cce8b4" exitCode=0 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.887563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" event={"ID":"94d68ac6-09be-4ca6-8221-f508337ec0c7","Type":"ContainerDied","Data":"7da83339064de66aef2ffeeb28a472af93299d9111a3a0d6ed18c271c2cce8b4"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.890016 4770 generic.go:334] "Generic (PLEG): container finished" podID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerID="028a69e195d308587c22bd9b31ad909606b0269f7c88819cd17e0a2731e821d3" exitCode=0 Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.890066 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerDied","Data":"028a69e195d308587c22bd9b31ad909606b0269f7c88819cd17e0a2731e821d3"} Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.896397 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.899496 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.990848 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:37:54 crc kubenswrapper[4770]: I1209 11:37:54.991458 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.003000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content\") pod \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.003096 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnwxz\" (UniqueName: \"kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz\") pod \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.003168 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities\") pod \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\" (UID: \"b9afd8a5-f75a-4aa5-8cba-a4699b25b847\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.004106 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities" (OuterVolumeSpecName: "utilities") pod "b9afd8a5-f75a-4aa5-8cba-a4699b25b847" (UID: "b9afd8a5-f75a-4aa5-8cba-a4699b25b847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.010438 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz" (OuterVolumeSpecName: "kube-api-access-nnwxz") pod "b9afd8a5-f75a-4aa5-8cba-a4699b25b847" (UID: "b9afd8a5-f75a-4aa5-8cba-a4699b25b847"). InnerVolumeSpecName "kube-api-access-nnwxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.015696 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.071194 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnsf7\" (UniqueName: \"kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7\") pod \"df7cd067-cdd9-42c3-a3eb-770391bd9250\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106642 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content\") pod \"df7cd067-cdd9-42c3-a3eb-770391bd9250\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106677 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities\") pod \"df7cd067-cdd9-42c3-a3eb-770391bd9250\" (UID: \"df7cd067-cdd9-42c3-a3eb-770391bd9250\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106741 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kwv8\" (UniqueName: \"kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8\") pod \"54352a0d-9e92-431b-9f78-6e9a721c760e\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106779 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content\") pod \"54352a0d-9e92-431b-9f78-6e9a721c760e\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.106806 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities\") pod \"54352a0d-9e92-431b-9f78-6e9a721c760e\" (UID: \"54352a0d-9e92-431b-9f78-6e9a721c760e\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.107060 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnwxz\" (UniqueName: \"kubernetes.io/projected/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-kube-api-access-nnwxz\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.107078 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.108076 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities" (OuterVolumeSpecName: "utilities") pod "54352a0d-9e92-431b-9f78-6e9a721c760e" (UID: "54352a0d-9e92-431b-9f78-6e9a721c760e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.111556 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities" (OuterVolumeSpecName: "utilities") pod "df7cd067-cdd9-42c3-a3eb-770391bd9250" (UID: "df7cd067-cdd9-42c3-a3eb-770391bd9250"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.120795 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8" (OuterVolumeSpecName: "kube-api-access-6kwv8") pod "54352a0d-9e92-431b-9f78-6e9a721c760e" (UID: "54352a0d-9e92-431b-9f78-6e9a721c760e"). InnerVolumeSpecName "kube-api-access-6kwv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.132407 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7" (OuterVolumeSpecName: "kube-api-access-lnsf7") pod "df7cd067-cdd9-42c3-a3eb-770391bd9250" (UID: "df7cd067-cdd9-42c3-a3eb-770391bd9250"). InnerVolumeSpecName "kube-api-access-lnsf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.132590 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df7cd067-cdd9-42c3-a3eb-770391bd9250" (UID: "df7cd067-cdd9-42c3-a3eb-770391bd9250"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.163462 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9afd8a5-f75a-4aa5-8cba-a4699b25b847" (UID: "b9afd8a5-f75a-4aa5-8cba-a4699b25b847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.208677 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tjhl\" (UniqueName: \"kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl\") pod \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.208795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics\") pod \"94d68ac6-09be-4ca6-8221-f508337ec0c7\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.208843 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9gvr\" (UniqueName: \"kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr\") pod \"94d68ac6-09be-4ca6-8221-f508337ec0c7\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.208878 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca\") pod \"94d68ac6-09be-4ca6-8221-f508337ec0c7\" (UID: \"94d68ac6-09be-4ca6-8221-f508337ec0c7\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.208996 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities\") pod \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209042 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content\") pod \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\" (UID: \"03311274-e8f5-4e5c-8dce-dce3e7235bfb\") " Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209343 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kwv8\" (UniqueName: \"kubernetes.io/projected/54352a0d-9e92-431b-9f78-6e9a721c760e-kube-api-access-6kwv8\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209356 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209370 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnsf7\" (UniqueName: \"kubernetes.io/projected/df7cd067-cdd9-42c3-a3eb-770391bd9250-kube-api-access-lnsf7\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209381 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9afd8a5-f75a-4aa5-8cba-a4699b25b847-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209390 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.209412 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df7cd067-cdd9-42c3-a3eb-770391bd9250-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.210275 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "94d68ac6-09be-4ca6-8221-f508337ec0c7" (UID: "94d68ac6-09be-4ca6-8221-f508337ec0c7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.212099 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl" (OuterVolumeSpecName: "kube-api-access-6tjhl") pod "03311274-e8f5-4e5c-8dce-dce3e7235bfb" (UID: "03311274-e8f5-4e5c-8dce-dce3e7235bfb"). InnerVolumeSpecName "kube-api-access-6tjhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.212211 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities" (OuterVolumeSpecName: "utilities") pod "03311274-e8f5-4e5c-8dce-dce3e7235bfb" (UID: "03311274-e8f5-4e5c-8dce-dce3e7235bfb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.213575 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "94d68ac6-09be-4ca6-8221-f508337ec0c7" (UID: "94d68ac6-09be-4ca6-8221-f508337ec0c7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.218645 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr" (OuterVolumeSpecName: "kube-api-access-j9gvr") pod "94d68ac6-09be-4ca6-8221-f508337ec0c7" (UID: "94d68ac6-09be-4ca6-8221-f508337ec0c7"). InnerVolumeSpecName "kube-api-access-j9gvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.280412 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03311274-e8f5-4e5c-8dce-dce3e7235bfb" (UID: "03311274-e8f5-4e5c-8dce-dce3e7235bfb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.293865 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54352a0d-9e92-431b-9f78-6e9a721c760e" (UID: "54352a0d-9e92-431b-9f78-6e9a721c760e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311102 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54352a0d-9e92-431b-9f78-6e9a721c760e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311164 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tjhl\" (UniqueName: \"kubernetes.io/projected/03311274-e8f5-4e5c-8dce-dce3e7235bfb-kube-api-access-6tjhl\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311179 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311190 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9gvr\" (UniqueName: \"kubernetes.io/projected/94d68ac6-09be-4ca6-8221-f508337ec0c7-kube-api-access-j9gvr\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311200 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94d68ac6-09be-4ca6-8221-f508337ec0c7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311213 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.311222 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03311274-e8f5-4e5c-8dce-dce3e7235bfb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.472393 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-twm9p"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.900024 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnc6l" event={"ID":"df7cd067-cdd9-42c3-a3eb-770391bd9250","Type":"ContainerDied","Data":"8d2769f08575d7e9211842622a722c686e89396ececad7b244701d03ee62a31b"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.900359 4770 scope.go:117] "RemoveContainer" containerID="32b37d1cd3dc61b78afc9fd2df735d736e635d9faa7feabada5417cb9180034c" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.900491 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnc6l" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.905589 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" event={"ID":"94d68ac6-09be-4ca6-8221-f508337ec0c7","Type":"ContainerDied","Data":"66972805ad510c9cf39be2451664b16c1c831da2b50c9a71e2968d642ca119b5"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.905810 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sfxhc" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.908111 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" event={"ID":"9e029e5b-b9b0-483a-ad49-29f8ceb43e61","Type":"ContainerStarted","Data":"736fd7f5cb302efd4808d7c82bf0e51e8c58d33939f5147d4751850f75402888"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.908184 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.908197 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" event={"ID":"9e029e5b-b9b0-483a-ad49-29f8ceb43e61","Type":"ContainerStarted","Data":"8e6ca4571db1ad3f4f16b1fbb61228751b1c0b9ef2823682e75ead74ad50cc97"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.909844 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-twm9p container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" start-of-body= Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.909885 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" podUID="9e029e5b-b9b0-483a-ad49-29f8ceb43e61" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.910085 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cng6d" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.910121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cng6d" event={"ID":"54352a0d-9e92-431b-9f78-6e9a721c760e","Type":"ContainerDied","Data":"1ba105edc28a244ec5dab91dabac077883dcb7b1a76dec29a587d6a2240ca5e9"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.917136 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84dm9" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.917786 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgrjv" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.918166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgrjv" event={"ID":"03311274-e8f5-4e5c-8dce-dce3e7235bfb","Type":"ContainerDied","Data":"35f296ed72881f77f4ee6c2272f5f9fea47f8cf1e4508405cdeb52faedd3fceb"} Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.928703 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.930153 4770 scope.go:117] "RemoveContainer" containerID="9a3f9a6d0bcdb2263b56bb42d75760009e85e9f2a219f015a4d43d28aa1b74f9" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.931640 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnc6l"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.946959 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.955045 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-84dm9"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.963322 4770 scope.go:117] "RemoveContainer" containerID="0488e723a757ad78544fe9badf548b399028787eef36d176bb2209c94f8b74d6" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.964128 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" podStartSLOduration=1.964107633 podStartE2EDuration="1.964107633s" podCreationTimestamp="2025-12-09 11:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:37:55.960773356 +0000 UTC m=+401.201531885" watchObservedRunningTime="2025-12-09 11:37:55.964107633 +0000 UTC m=+401.204866152" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.980426 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.988566 4770 scope.go:117] "RemoveContainer" containerID="7da83339064de66aef2ffeeb28a472af93299d9111a3a0d6ed18c271c2cce8b4" Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.989208 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cng6d"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.994639 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:37:55 crc kubenswrapper[4770]: I1209 11:37:55.997721 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgrjv"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.008086 4770 scope.go:117] "RemoveContainer" containerID="028a69e195d308587c22bd9b31ad909606b0269f7c88819cd17e0a2731e821d3" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.010866 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.015190 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sfxhc"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.027445 4770 scope.go:117] "RemoveContainer" containerID="b3ba907dea74f1a3f20b73a5b54bc7ac0e87a5b711e183d72bc4e9d3dff3c97a" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.057887 4770 scope.go:117] "RemoveContainer" containerID="1dddc829d6c14f034293742bc30b0847e27a62076a5e426e755b094f9ac9c287" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.088347 4770 scope.go:117] "RemoveContainer" containerID="7a0bfea976d7b11910250169c1146afa4ec8aaaa7a9d9baf008fad3ceb4c65a7" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.114264 4770 scope.go:117] "RemoveContainer" containerID="39931cd47ae9333df0b2c341c3727f2459ff21f87e4a4c9456c0b6bdc97a3f79" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.137248 4770 scope.go:117] "RemoveContainer" containerID="73f827ed94b8021ade3fc948aeb05e0a079cf4cd524778b88262ea81959314e0" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620198 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qfsls"] Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620512 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620534 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620550 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620558 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620571 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620581 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620595 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620603 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620615 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620623 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620631 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620639 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620652 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620659 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620669 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620676 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620688 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620695 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620705 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620712 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620726 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620732 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="extract-utilities" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620744 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620753 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="extract-content" Dec 09 11:37:56 crc kubenswrapper[4770]: E1209 11:37:56.620761 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620768 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620873 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" containerName="marketplace-operator" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620888 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620913 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620927 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.620936 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" containerName="registry-server" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.621894 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.626873 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.640674 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfsls"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.649149 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-utilities\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.649210 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-catalog-content\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.649242 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvkdz\" (UniqueName: \"kubernetes.io/projected/d42dfe0d-819e-48a0-b808-e2a34f531355-kube-api-access-nvkdz\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.750930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-utilities\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.751031 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-catalog-content\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.751069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvkdz\" (UniqueName: \"kubernetes.io/projected/d42dfe0d-819e-48a0-b808-e2a34f531355-kube-api-access-nvkdz\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.752509 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-utilities\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.752767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42dfe0d-819e-48a0-b808-e2a34f531355-catalog-content\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.773370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvkdz\" (UniqueName: \"kubernetes.io/projected/d42dfe0d-819e-48a0-b808-e2a34f531355-kube-api-access-nvkdz\") pod \"redhat-marketplace-qfsls\" (UID: \"d42dfe0d-819e-48a0-b808-e2a34f531355\") " pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.820734 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vn6tj"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.821675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.824832 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.852418 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vn6tj"] Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.852862 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-catalog-content\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.852980 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zswgf\" (UniqueName: \"kubernetes.io/projected/0111570a-aa10-4ffb-a876-9b24fcc9415a-kube-api-access-zswgf\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.853018 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-utilities\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.931013 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-twm9p" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.943043 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.958661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-catalog-content\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.958853 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zswgf\" (UniqueName: \"kubernetes.io/projected/0111570a-aa10-4ffb-a876-9b24fcc9415a-kube-api-access-zswgf\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.958917 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-utilities\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.961256 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-utilities\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.962666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0111570a-aa10-4ffb-a876-9b24fcc9415a-catalog-content\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:56 crc kubenswrapper[4770]: I1209 11:37:56.982971 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zswgf\" (UniqueName: \"kubernetes.io/projected/0111570a-aa10-4ffb-a876-9b24fcc9415a-kube-api-access-zswgf\") pod \"certified-operators-vn6tj\" (UID: \"0111570a-aa10-4ffb-a876-9b24fcc9415a\") " pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.150610 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.213459 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfsls"] Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.353446 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03311274-e8f5-4e5c-8dce-dce3e7235bfb" path="/var/lib/kubelet/pods/03311274-e8f5-4e5c-8dce-dce3e7235bfb/volumes" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.354874 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54352a0d-9e92-431b-9f78-6e9a721c760e" path="/var/lib/kubelet/pods/54352a0d-9e92-431b-9f78-6e9a721c760e/volumes" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.355750 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d68ac6-09be-4ca6-8221-f508337ec0c7" path="/var/lib/kubelet/pods/94d68ac6-09be-4ca6-8221-f508337ec0c7/volumes" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.359590 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9afd8a5-f75a-4aa5-8cba-a4699b25b847" path="/var/lib/kubelet/pods/b9afd8a5-f75a-4aa5-8cba-a4699b25b847/volumes" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.361228 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df7cd067-cdd9-42c3-a3eb-770391bd9250" path="/var/lib/kubelet/pods/df7cd067-cdd9-42c3-a3eb-770391bd9250/volumes" Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.362085 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vn6tj"] Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.937082 4770 generic.go:334] "Generic (PLEG): container finished" podID="d42dfe0d-819e-48a0-b808-e2a34f531355" containerID="b33d4ca2d487a12732f67a59f3dd1cc1768fa297f0c9c72736518c21f74ad508" exitCode=0 Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.937207 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfsls" event={"ID":"d42dfe0d-819e-48a0-b808-e2a34f531355","Type":"ContainerDied","Data":"b33d4ca2d487a12732f67a59f3dd1cc1768fa297f0c9c72736518c21f74ad508"} Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.937275 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfsls" event={"ID":"d42dfe0d-819e-48a0-b808-e2a34f531355","Type":"ContainerStarted","Data":"c7b924d9378d1613e6a85788fa64133d5b5a5129aadceb894392df0420c86b51"} Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.939501 4770 generic.go:334] "Generic (PLEG): container finished" podID="0111570a-aa10-4ffb-a876-9b24fcc9415a" containerID="098c6f6cb1badb4cafe59915bccbecc4e504b804180c17afe2b943dfce63538d" exitCode=0 Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.940263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vn6tj" event={"ID":"0111570a-aa10-4ffb-a876-9b24fcc9415a","Type":"ContainerDied","Data":"098c6f6cb1badb4cafe59915bccbecc4e504b804180c17afe2b943dfce63538d"} Dec 09 11:37:57 crc kubenswrapper[4770]: I1209 11:37:57.940369 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vn6tj" event={"ID":"0111570a-aa10-4ffb-a876-9b24fcc9415a","Type":"ContainerStarted","Data":"d10fabff2a7c640448bf6c10faed6b6554dfa8e028fb70e0eeceec21880926f4"} Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.020110 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.021969 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.025336 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.034121 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.192633 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.192733 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r585\" (UniqueName: \"kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.193271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.217011 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2b6tg"] Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.218595 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.224201 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.232603 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2b6tg"] Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.295393 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.295488 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r585\" (UniqueName: \"kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.295543 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.296317 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.296522 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.327464 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r585\" (UniqueName: \"kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585\") pod \"redhat-operators-twvqx\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.339757 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.396954 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-catalog-content\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.397018 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4hwl\" (UniqueName: \"kubernetes.io/projected/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-kube-api-access-z4hwl\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.397090 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-utilities\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.499166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-utilities\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.500307 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-catalog-content\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.500440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4hwl\" (UniqueName: \"kubernetes.io/projected/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-kube-api-access-z4hwl\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.500598 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-catalog-content\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.499797 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-utilities\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.524049 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4hwl\" (UniqueName: \"kubernetes.io/projected/ac16f7d6-2737-46df-8cf8-6df5fafbb9c6-kube-api-access-z4hwl\") pod \"community-operators-2b6tg\" (UID: \"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6\") " pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:37:59 crc kubenswrapper[4770]: I1209 11:37:59.542826 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.246355 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 11:38:00 crc kubenswrapper[4770]: W1209 11:38:00.257496 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7ea1b87_f78d_4999_a66a_3ca61f520d42.slice/crio-ad284524dbb643e747b7cbe6283a57f0fceb6b6a61f46c6111c3e5b5a18d5fc4 WatchSource:0}: Error finding container ad284524dbb643e747b7cbe6283a57f0fceb6b6a61f46c6111c3e5b5a18d5fc4: Status 404 returned error can't find the container with id ad284524dbb643e747b7cbe6283a57f0fceb6b6a61f46c6111c3e5b5a18d5fc4 Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.352594 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2b6tg"] Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.974735 4770 generic.go:334] "Generic (PLEG): container finished" podID="0111570a-aa10-4ffb-a876-9b24fcc9415a" containerID="df5a24ee46f7dd04405afbddf7b03261e33a896288b4c4108f77c54c08c7dc9e" exitCode=0 Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.974795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vn6tj" event={"ID":"0111570a-aa10-4ffb-a876-9b24fcc9415a","Type":"ContainerDied","Data":"df5a24ee46f7dd04405afbddf7b03261e33a896288b4c4108f77c54c08c7dc9e"} Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.980352 4770 generic.go:334] "Generic (PLEG): container finished" podID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerID="c64d6bc91e0ec2cb3b9d127bd55fb64eab1fac535d0c0c43ce87c2958762b139" exitCode=0 Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.982405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerDied","Data":"c64d6bc91e0ec2cb3b9d127bd55fb64eab1fac535d0c0c43ce87c2958762b139"} Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.982523 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerStarted","Data":"ad284524dbb643e747b7cbe6283a57f0fceb6b6a61f46c6111c3e5b5a18d5fc4"} Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.984364 4770 generic.go:334] "Generic (PLEG): container finished" podID="ac16f7d6-2737-46df-8cf8-6df5fafbb9c6" containerID="531ec1caf1eb6e2136cf06f6a1d80e9005df1f6bc47fd55bfd66d6b9321e71fa" exitCode=0 Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.984436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b6tg" event={"ID":"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6","Type":"ContainerDied","Data":"531ec1caf1eb6e2136cf06f6a1d80e9005df1f6bc47fd55bfd66d6b9321e71fa"} Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.984454 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b6tg" event={"ID":"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6","Type":"ContainerStarted","Data":"6a76e407181d6723d78897c5c968a40a3bf2f6fa2e6cb02554e305af7790ad7d"} Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.994319 4770 generic.go:334] "Generic (PLEG): container finished" podID="d42dfe0d-819e-48a0-b808-e2a34f531355" containerID="28b3fd7d907798ddc52908ba4cc059f3144287a1e4aa579babe3bede60d23657" exitCode=0 Dec 09 11:38:00 crc kubenswrapper[4770]: I1209 11:38:00.994369 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfsls" event={"ID":"d42dfe0d-819e-48a0-b808-e2a34f531355","Type":"ContainerDied","Data":"28b3fd7d907798ddc52908ba4cc059f3144287a1e4aa579babe3bede60d23657"} Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.008710 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerStarted","Data":"7662d5b337f050813639e1ae72957acb9ad7eca1658c464f58447e4ec7d8a8dc"} Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.012153 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b6tg" event={"ID":"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6","Type":"ContainerStarted","Data":"427bcbcebf2ad5cae51e2b01a2d2899ff092937333e48c458d0e251eaef7f0d8"} Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.015279 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfsls" event={"ID":"d42dfe0d-819e-48a0-b808-e2a34f531355","Type":"ContainerStarted","Data":"6bc083f3afd693b3eeedc736df7e98455a93e80196cab147ba6e3d2b85089711"} Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.018346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vn6tj" event={"ID":"0111570a-aa10-4ffb-a876-9b24fcc9415a","Type":"ContainerStarted","Data":"cadbfdaec97442a1b32dfd2b1a2b126e1ed03add789467d96ee31495b157f939"} Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.054877 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qfsls" podStartSLOduration=2.521773363 podStartE2EDuration="6.054854567s" podCreationTimestamp="2025-12-09 11:37:56 +0000 UTC" firstStartedPulling="2025-12-09 11:37:57.940891596 +0000 UTC m=+403.181650115" lastFinishedPulling="2025-12-09 11:38:01.4739728 +0000 UTC m=+406.714731319" observedRunningTime="2025-12-09 11:38:02.053623905 +0000 UTC m=+407.294382424" watchObservedRunningTime="2025-12-09 11:38:02.054854567 +0000 UTC m=+407.295613086" Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.076854 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vn6tj" podStartSLOduration=2.623649459 podStartE2EDuration="6.076829118s" podCreationTimestamp="2025-12-09 11:37:56 +0000 UTC" firstStartedPulling="2025-12-09 11:37:57.942361964 +0000 UTC m=+403.183120483" lastFinishedPulling="2025-12-09 11:38:01.395541623 +0000 UTC m=+406.636300142" observedRunningTime="2025-12-09 11:38:02.07536838 +0000 UTC m=+407.316126899" watchObservedRunningTime="2025-12-09 11:38:02.076829118 +0000 UTC m=+407.317587637" Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.473589 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:38:02 crc kubenswrapper[4770]: I1209 11:38:02.473670 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:38:03 crc kubenswrapper[4770]: I1209 11:38:03.028275 4770 generic.go:334] "Generic (PLEG): container finished" podID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerID="7662d5b337f050813639e1ae72957acb9ad7eca1658c464f58447e4ec7d8a8dc" exitCode=0 Dec 09 11:38:03 crc kubenswrapper[4770]: I1209 11:38:03.028402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerDied","Data":"7662d5b337f050813639e1ae72957acb9ad7eca1658c464f58447e4ec7d8a8dc"} Dec 09 11:38:03 crc kubenswrapper[4770]: I1209 11:38:03.033283 4770 generic.go:334] "Generic (PLEG): container finished" podID="ac16f7d6-2737-46df-8cf8-6df5fafbb9c6" containerID="427bcbcebf2ad5cae51e2b01a2d2899ff092937333e48c458d0e251eaef7f0d8" exitCode=0 Dec 09 11:38:03 crc kubenswrapper[4770]: I1209 11:38:03.034289 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b6tg" event={"ID":"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6","Type":"ContainerDied","Data":"427bcbcebf2ad5cae51e2b01a2d2899ff092937333e48c458d0e251eaef7f0d8"} Dec 09 11:38:04 crc kubenswrapper[4770]: I1209 11:38:04.042722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerStarted","Data":"d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de"} Dec 09 11:38:04 crc kubenswrapper[4770]: I1209 11:38:04.045861 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b6tg" event={"ID":"ac16f7d6-2737-46df-8cf8-6df5fafbb9c6","Type":"ContainerStarted","Data":"5259990c2b1071bbee5ae6c4a5f798bf4b79147c4b1d8f1475b63b8a5c7f47e4"} Dec 09 11:38:04 crc kubenswrapper[4770]: I1209 11:38:04.067813 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-twvqx" podStartSLOduration=2.578768732 podStartE2EDuration="5.067791188s" podCreationTimestamp="2025-12-09 11:37:59 +0000 UTC" firstStartedPulling="2025-12-09 11:38:00.982529506 +0000 UTC m=+406.223288025" lastFinishedPulling="2025-12-09 11:38:03.471551962 +0000 UTC m=+408.712310481" observedRunningTime="2025-12-09 11:38:04.065961851 +0000 UTC m=+409.306720360" watchObservedRunningTime="2025-12-09 11:38:04.067791188 +0000 UTC m=+409.308549707" Dec 09 11:38:04 crc kubenswrapper[4770]: I1209 11:38:04.088920 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2b6tg" podStartSLOduration=2.572030387 podStartE2EDuration="5.088866836s" podCreationTimestamp="2025-12-09 11:37:59 +0000 UTC" firstStartedPulling="2025-12-09 11:38:00.992825833 +0000 UTC m=+406.233584352" lastFinishedPulling="2025-12-09 11:38:03.509662282 +0000 UTC m=+408.750420801" observedRunningTime="2025-12-09 11:38:04.084733568 +0000 UTC m=+409.325492087" watchObservedRunningTime="2025-12-09 11:38:04.088866836 +0000 UTC m=+409.329625355" Dec 09 11:38:06 crc kubenswrapper[4770]: I1209 11:38:06.943740 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:38:06 crc kubenswrapper[4770]: I1209 11:38:06.948861 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:38:07 crc kubenswrapper[4770]: I1209 11:38:07.002359 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:38:07 crc kubenswrapper[4770]: I1209 11:38:07.105692 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qfsls" Dec 09 11:38:07 crc kubenswrapper[4770]: I1209 11:38:07.151018 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:38:07 crc kubenswrapper[4770]: I1209 11:38:07.151100 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:38:07 crc kubenswrapper[4770]: I1209 11:38:07.198995 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:38:08 crc kubenswrapper[4770]: I1209 11:38:08.123292 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vn6tj" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.348836 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.349424 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.387382 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.544044 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.544138 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:38:09 crc kubenswrapper[4770]: I1209 11:38:09.592066 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:38:10 crc kubenswrapper[4770]: I1209 11:38:10.127046 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2b6tg" Dec 09 11:38:10 crc kubenswrapper[4770]: I1209 11:38:10.130317 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 11:38:32 crc kubenswrapper[4770]: I1209 11:38:32.473412 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:38:32 crc kubenswrapper[4770]: I1209 11:38:32.474181 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:38:32 crc kubenswrapper[4770]: I1209 11:38:32.474255 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:38:32 crc kubenswrapper[4770]: I1209 11:38:32.475214 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:38:32 crc kubenswrapper[4770]: I1209 11:38:32.475314 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835" gracePeriod=600 Dec 09 11:38:33 crc kubenswrapper[4770]: I1209 11:38:33.229311 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835" exitCode=0 Dec 09 11:38:33 crc kubenswrapper[4770]: I1209 11:38:33.229413 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835"} Dec 09 11:38:33 crc kubenswrapper[4770]: I1209 11:38:33.229659 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12"} Dec 09 11:38:33 crc kubenswrapper[4770]: I1209 11:38:33.229700 4770 scope.go:117] "RemoveContainer" containerID="5fa6a611c0ab1253dc628f7b037866acd675719b11a854256b99241d07fe6f29" Dec 09 11:40:15 crc kubenswrapper[4770]: I1209 11:40:15.672965 4770 scope.go:117] "RemoveContainer" containerID="e46abe21ea5e58e0dbd03e1ba614f27b827ddd2d5a74285f928492ba485e9dba" Dec 09 11:40:32 crc kubenswrapper[4770]: I1209 11:40:32.474034 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:40:32 crc kubenswrapper[4770]: I1209 11:40:32.474631 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:41:02 crc kubenswrapper[4770]: I1209 11:41:02.474201 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:41:02 crc kubenswrapper[4770]: I1209 11:41:02.474802 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:41:32 crc kubenswrapper[4770]: I1209 11:41:32.474108 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:41:32 crc kubenswrapper[4770]: I1209 11:41:32.474651 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:41:32 crc kubenswrapper[4770]: I1209 11:41:32.474708 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:41:32 crc kubenswrapper[4770]: I1209 11:41:32.475317 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:41:32 crc kubenswrapper[4770]: I1209 11:41:32.475370 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12" gracePeriod=600 Dec 09 11:41:33 crc kubenswrapper[4770]: I1209 11:41:33.307761 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12" exitCode=0 Dec 09 11:41:33 crc kubenswrapper[4770]: I1209 11:41:33.307846 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12"} Dec 09 11:41:33 crc kubenswrapper[4770]: I1209 11:41:33.308566 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b"} Dec 09 11:41:33 crc kubenswrapper[4770]: I1209 11:41:33.308602 4770 scope.go:117] "RemoveContainer" containerID="cf645011fff4594477d21ae1b27bfd2057da5b393fbe21cd6a82425553fdb835" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.022718 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qpt8q"] Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.024595 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.042366 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qpt8q"] Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.128879 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0b6afa50-26da-4e30-8a72-3cc627e323f9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.128962 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-bound-sa-token\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.128990 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-trusted-ca\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.129041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.129249 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-tls\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.129308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pk6n\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-kube-api-access-9pk6n\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.129342 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-certificates\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.129519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0b6afa50-26da-4e30-8a72-3cc627e323f9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.158065 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231374 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-tls\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231433 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pk6n\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-kube-api-access-9pk6n\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231463 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-certificates\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231523 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0b6afa50-26da-4e30-8a72-3cc627e323f9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0b6afa50-26da-4e30-8a72-3cc627e323f9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231573 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-bound-sa-token\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.231601 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-trusted-ca\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.232346 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0b6afa50-26da-4e30-8a72-3cc627e323f9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.233232 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-trusted-ca\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.233241 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-certificates\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.242033 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-registry-tls\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.242142 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0b6afa50-26da-4e30-8a72-3cc627e323f9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.251296 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pk6n\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-kube-api-access-9pk6n\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.256149 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b6afa50-26da-4e30-8a72-3cc627e323f9-bound-sa-token\") pod \"image-registry-66df7c8f76-qpt8q\" (UID: \"0b6afa50-26da-4e30-8a72-3cc627e323f9\") " pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.343487 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:52 crc kubenswrapper[4770]: I1209 11:41:52.543739 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qpt8q"] Dec 09 11:41:53 crc kubenswrapper[4770]: I1209 11:41:53.426472 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" event={"ID":"0b6afa50-26da-4e30-8a72-3cc627e323f9","Type":"ContainerStarted","Data":"866edec0be1f28b3ef6a17a5dcc8dfd0eb5684fdd96343ac31265e7ccd916e9c"} Dec 09 11:41:53 crc kubenswrapper[4770]: I1209 11:41:53.426807 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" event={"ID":"0b6afa50-26da-4e30-8a72-3cc627e323f9","Type":"ContainerStarted","Data":"f0b25a20a7a6900123c81ca168fc753ff880f25fc838ed63e5883217668212bf"} Dec 09 11:41:53 crc kubenswrapper[4770]: I1209 11:41:53.426825 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:41:53 crc kubenswrapper[4770]: I1209 11:41:53.447110 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" podStartSLOduration=2.447061438 podStartE2EDuration="2.447061438s" podCreationTimestamp="2025-12-09 11:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:41:53.442984243 +0000 UTC m=+638.683742772" watchObservedRunningTime="2025-12-09 11:41:53.447061438 +0000 UTC m=+638.687819967" Dec 09 11:42:12 crc kubenswrapper[4770]: I1209 11:42:12.352451 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-qpt8q" Dec 09 11:42:12 crc kubenswrapper[4770]: I1209 11:42:12.416572 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:42:15 crc kubenswrapper[4770]: I1209 11:42:15.719216 4770 scope.go:117] "RemoveContainer" containerID="9186eca8ab5b6b6f088b79a4bedaf5039e7602d97c2cfeb8e736335bf0ff68f9" Dec 09 11:42:15 crc kubenswrapper[4770]: I1209 11:42:15.747294 4770 scope.go:117] "RemoveContainer" containerID="b3d670f2a8db28e2d2ed26e3ca1e3a55cf7e179cc66f51b44f0fa7b982088190" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.476123 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" podUID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" containerName="registry" containerID="cri-o://96dca8ae8546027ce121a9bf5912cbb03315e67eff86b61f3c65872a42d8f069" gracePeriod=30 Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.688174 4770 generic.go:334] "Generic (PLEG): container finished" podID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" containerID="96dca8ae8546027ce121a9bf5912cbb03315e67eff86b61f3c65872a42d8f069" exitCode=0 Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.688265 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" event={"ID":"bd7906e7-dbf7-4668-aec9-8e9b778c9452","Type":"ContainerDied","Data":"96dca8ae8546027ce121a9bf5912cbb03315e67eff86b61f3c65872a42d8f069"} Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.804030 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.936927 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.936997 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937035 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937239 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937284 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937326 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937388 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lww2l\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.937429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token\") pod \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\" (UID: \"bd7906e7-dbf7-4668-aec9-8e9b778c9452\") " Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.938470 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.938734 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.946397 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l" (OuterVolumeSpecName: "kube-api-access-lww2l") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "kube-api-access-lww2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.946692 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.949426 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.948411 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.949866 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:42:37 crc kubenswrapper[4770]: I1209 11:42:37.955818 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bd7906e7-dbf7-4668-aec9-8e9b778c9452" (UID: "bd7906e7-dbf7-4668-aec9-8e9b778c9452"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039190 4770 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd7906e7-dbf7-4668-aec9-8e9b778c9452-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039229 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039243 4770 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039257 4770 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd7906e7-dbf7-4668-aec9-8e9b778c9452-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039268 4770 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039279 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lww2l\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-kube-api-access-lww2l\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.039288 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd7906e7-dbf7-4668-aec9-8e9b778c9452-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.695652 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" event={"ID":"bd7906e7-dbf7-4668-aec9-8e9b778c9452","Type":"ContainerDied","Data":"612e39b1bd3c342112009aec75a72feb29e7afbc75db6bcfb6f80d88b0224291"} Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.695711 4770 scope.go:117] "RemoveContainer" containerID="96dca8ae8546027ce121a9bf5912cbb03315e67eff86b61f3c65872a42d8f069" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.695740 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gxm6f" Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.736584 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:42:38 crc kubenswrapper[4770]: I1209 11:42:38.742106 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gxm6f"] Dec 09 11:42:39 crc kubenswrapper[4770]: I1209 11:42:39.351786 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" path="/var/lib/kubelet/pods/bd7906e7-dbf7-4668-aec9-8e9b778c9452/volumes" Dec 09 11:43:32 crc kubenswrapper[4770]: I1209 11:43:32.474553 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:43:32 crc kubenswrapper[4770]: I1209 11:43:32.476043 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:44:02 crc kubenswrapper[4770]: I1209 11:44:02.473422 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:44:02 crc kubenswrapper[4770]: I1209 11:44:02.473962 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:44:11 crc kubenswrapper[4770]: I1209 11:44:11.925590 4770 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 11:44:32 crc kubenswrapper[4770]: I1209 11:44:32.473955 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:44:32 crc kubenswrapper[4770]: I1209 11:44:32.475016 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:44:32 crc kubenswrapper[4770]: I1209 11:44:32.475127 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:44:32 crc kubenswrapper[4770]: I1209 11:44:32.476055 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:44:32 crc kubenswrapper[4770]: I1209 11:44:32.476121 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b" gracePeriod=600 Dec 09 11:44:33 crc kubenswrapper[4770]: I1209 11:44:33.370116 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b" exitCode=0 Dec 09 11:44:33 crc kubenswrapper[4770]: I1209 11:44:33.370188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b"} Dec 09 11:44:33 crc kubenswrapper[4770]: I1209 11:44:33.370788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24"} Dec 09 11:44:33 crc kubenswrapper[4770]: I1209 11:44:33.370814 4770 scope.go:117] "RemoveContainer" containerID="01c3132cade9346828360f55b78cfa6f2bc65e87f5589218238f0ecf02107b12" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.259761 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zt2b2"] Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261058 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-controller" containerID="cri-o://6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261205 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="northd" containerID="cri-o://37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261197 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="nbdb" containerID="cri-o://416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261278 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-node" containerID="cri-o://11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261324 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-acl-logging" containerID="cri-o://77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261262 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.261748 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="sbdb" containerID="cri-o://79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.308541 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" containerID="cri-o://33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" gracePeriod=30 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.415624 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/2.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.416985 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/1.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.417047 4770 generic.go:334] "Generic (PLEG): container finished" podID="4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86" containerID="55d7d7cac78fc0627d84150710ccaaeb7dba76a2a111e5f3e2b56f76be78c372" exitCode=2 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.417143 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerDied","Data":"55d7d7cac78fc0627d84150710ccaaeb7dba76a2a111e5f3e2b56f76be78c372"} Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.417191 4770 scope.go:117] "RemoveContainer" containerID="b8c5923659a25ede5e5103bb7d8d5c944fc2d09fba3f4f2ec261dbc667abfb59" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.417783 4770 scope.go:117] "RemoveContainer" containerID="55d7d7cac78fc0627d84150710ccaaeb7dba76a2a111e5f3e2b56f76be78c372" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.424157 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.426803 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-acl-logging/0.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.427672 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-controller/0.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.428233 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" exitCode=0 Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.428353 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.658053 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.661152 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-acl-logging/0.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.662538 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-controller/0.log" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.663466 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749547 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nshst"] Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749830 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="sbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749849 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="sbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749862 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="nbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749871 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="nbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749880 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749888 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749921 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kubecfg-setup" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749932 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kubecfg-setup" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749946 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749954 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749966 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-node" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749974 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-node" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.749984 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" containerName="registry" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.749992 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" containerName="registry" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750003 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750011 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750022 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750030 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750046 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="northd" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750054 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="northd" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750065 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750073 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750085 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-acl-logging" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750095 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-acl-logging" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750105 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750114 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750239 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-node" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750254 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750263 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="northd" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750272 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750283 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750292 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750300 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="sbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750309 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750317 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="nbdb" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750325 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-acl-logging" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750335 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovn-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750345 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7906e7-dbf7-4668-aec9-8e9b778c9452" containerName="registry" Dec 09 11:44:38 crc kubenswrapper[4770]: E1209 11:44:38.750462 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750472 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.750597 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerName="ovnkube-controller" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.752716 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795515 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795584 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795620 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795705 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795786 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795887 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795938 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.795969 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796011 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796013 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796103 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash" (OuterVolumeSpecName: "host-slash") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796111 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796164 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796194 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796215 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796260 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796278 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796306 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmjvf\" (UniqueName: \"kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf\") pod \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\" (UID: \"eaa774d4-1a7d-4731-a47b-d1d97c88869e\") " Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796658 4770 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-slash\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796676 4770 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796686 4770 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.796857 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797041 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797046 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797091 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797121 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log" (OuterVolumeSpecName: "node-log") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797121 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797149 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797178 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797208 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797363 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797439 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket" (OuterVolumeSpecName: "log-socket") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.797583 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.798322 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.808465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf" (OuterVolumeSpecName: "kube-api-access-xmjvf") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "kube-api-access-xmjvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.824260 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.830281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "eaa774d4-1a7d-4731-a47b-d1d97c88869e" (UID: "eaa774d4-1a7d-4731-a47b-d1d97c88869e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897370 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-bin\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897434 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-systemd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-netd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897864 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-env-overrides\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897945 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-log-socket\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.897982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-kubelet\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898041 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzzxg\" (UniqueName: \"kubernetes.io/projected/ceb5b168-5d89-413c-b590-0b29d1a60b29-kube-api-access-nzzxg\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898101 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-slash\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898129 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-var-lib-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898155 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-node-log\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898173 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-config\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898211 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-netns\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898231 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898267 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-etc-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898324 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898358 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-script-lib\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898427 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-systemd-units\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-ovn\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898499 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovn-node-metrics-cert\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898587 4770 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898607 4770 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898619 4770 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-node-log\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898629 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898642 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898653 4770 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898663 4770 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898672 4770 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-log-socket\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898681 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa774d4-1a7d-4731-a47b-d1d97c88869e-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898691 4770 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898700 4770 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898710 4770 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898724 4770 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898732 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmjvf\" (UniqueName: \"kubernetes.io/projected/eaa774d4-1a7d-4731-a47b-d1d97c88869e-kube-api-access-xmjvf\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898743 4770 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898752 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa774d4-1a7d-4731-a47b-d1d97c88869e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:38 crc kubenswrapper[4770]: I1209 11:44:38.898762 4770 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaa774d4-1a7d-4731-a47b-d1d97c88869e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000396 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-etc-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000468 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000492 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-script-lib\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000534 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-systemd-units\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000562 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-ovn\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovn-node-metrics-cert\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-bin\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000628 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-systemd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000645 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000664 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-netd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-env-overrides\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000703 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-log-socket\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000726 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-kubelet\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000747 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzzxg\" (UniqueName: \"kubernetes.io/projected/ceb5b168-5d89-413c-b590-0b29d1a60b29-kube-api-access-nzzxg\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000773 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-slash\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000792 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-var-lib-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-node-log\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000824 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-config\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000849 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-netns\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000868 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.000969 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.001016 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-etc-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.001042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.001836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-script-lib\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.001884 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-systemd-units\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.001925 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-ovn\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.002490 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-kubelet\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.002538 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-bin\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.002568 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-systemd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.002591 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-run-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.002619 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-cni-netd\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-env-overrides\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003087 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-log-socket\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003116 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-var-lib-openvswitch\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003186 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-slash\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003227 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-host-run-netns\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.003254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb5b168-5d89-413c-b590-0b29d1a60b29-node-log\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.004218 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovnkube-config\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.006557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb5b168-5d89-413c-b590-0b29d1a60b29-ovn-node-metrics-cert\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.021409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzzxg\" (UniqueName: \"kubernetes.io/projected/ceb5b168-5d89-413c-b590-0b29d1a60b29-kube-api-access-nzzxg\") pod \"ovnkube-node-nshst\" (UID: \"ceb5b168-5d89-413c-b590-0b29d1a60b29\") " pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.071719 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:39 crc kubenswrapper[4770]: W1209 11:44:39.093374 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podceb5b168_5d89_413c_b590_0b29d1a60b29.slice/crio-dcc637178b53656d3b2a454ec8572cbc2349b452cce952b2085cf045ef052ccd WatchSource:0}: Error finding container dcc637178b53656d3b2a454ec8572cbc2349b452cce952b2085cf045ef052ccd: Status 404 returned error can't find the container with id dcc637178b53656d3b2a454ec8572cbc2349b452cce952b2085cf045ef052ccd Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.447505 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-q2lxs_4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86/kube-multus/2.log" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.448040 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-q2lxs" event={"ID":"4eb284c3-ad1f-44f7-9bf7-f1a477d4dc86","Type":"ContainerStarted","Data":"cf40e1306b09580425c1b99b7bb180332a21a96577bd3e6fdbbe51e91607ba1c"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.452692 4770 generic.go:334] "Generic (PLEG): container finished" podID="ceb5b168-5d89-413c-b590-0b29d1a60b29" containerID="c2d9df6ccf1849fcf7e75d9ef84cd5d78931807ada3e28ee2652d3576c8e1096" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.452818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerDied","Data":"c2d9df6ccf1849fcf7e75d9ef84cd5d78931807ada3e28ee2652d3576c8e1096"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.452865 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"dcc637178b53656d3b2a454ec8572cbc2349b452cce952b2085cf045ef052ccd"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.456535 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovnkube-controller/3.log" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.461032 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-acl-logging/0.log" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.461623 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zt2b2_eaa774d4-1a7d-4731-a47b-d1d97c88869e/ovn-controller/0.log" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.461984 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462017 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462030 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462042 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462052 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" exitCode=0 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462061 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" exitCode=143 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462074 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" exitCode=143 Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462117 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462164 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462179 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462208 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462225 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462241 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462257 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462264 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462272 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462279 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462286 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462294 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462303 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462311 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462334 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462347 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462355 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462362 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462371 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462378 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462386 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462393 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462400 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462407 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" event={"ID":"eaa774d4-1a7d-4731-a47b-d1d97c88869e","Type":"ContainerDied","Data":"8668c6db0a39095b91845c2a1d06c8f8cab407f6c46944fc5539a7c9721fce2e"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462440 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462449 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462457 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462464 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462472 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462480 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462489 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462497 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462504 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462511 4770 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462334 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.462290 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zt2b2" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.491300 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.535883 4770 scope.go:117] "RemoveContainer" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.536106 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zt2b2"] Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.538707 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zt2b2"] Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.564409 4770 scope.go:117] "RemoveContainer" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.597723 4770 scope.go:117] "RemoveContainer" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.615648 4770 scope.go:117] "RemoveContainer" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.649035 4770 scope.go:117] "RemoveContainer" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.669839 4770 scope.go:117] "RemoveContainer" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.700492 4770 scope.go:117] "RemoveContainer" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.726599 4770 scope.go:117] "RemoveContainer" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.761084 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.761618 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.761675 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} err="failed to get container status \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.761714 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.762219 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": container with ID starting with 395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b not found: ID does not exist" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.762247 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} err="failed to get container status \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": rpc error: code = NotFound desc = could not find container \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": container with ID starting with 395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.762274 4770 scope.go:117] "RemoveContainer" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.762664 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": container with ID starting with 79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41 not found: ID does not exist" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.762698 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} err="failed to get container status \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": rpc error: code = NotFound desc = could not find container \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": container with ID starting with 79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.762723 4770 scope.go:117] "RemoveContainer" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.763250 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": container with ID starting with 416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8 not found: ID does not exist" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.763275 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} err="failed to get container status \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": rpc error: code = NotFound desc = could not find container \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": container with ID starting with 416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.763293 4770 scope.go:117] "RemoveContainer" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.763531 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": container with ID starting with 37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6 not found: ID does not exist" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.763554 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} err="failed to get container status \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": rpc error: code = NotFound desc = could not find container \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": container with ID starting with 37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.763569 4770 scope.go:117] "RemoveContainer" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.764073 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": container with ID starting with 1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e not found: ID does not exist" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764098 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} err="failed to get container status \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": rpc error: code = NotFound desc = could not find container \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": container with ID starting with 1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764119 4770 scope.go:117] "RemoveContainer" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.764409 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": container with ID starting with 11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b not found: ID does not exist" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764432 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} err="failed to get container status \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": rpc error: code = NotFound desc = could not find container \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": container with ID starting with 11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764448 4770 scope.go:117] "RemoveContainer" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.764690 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": container with ID starting with 77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5 not found: ID does not exist" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764716 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} err="failed to get container status \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": rpc error: code = NotFound desc = could not find container \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": container with ID starting with 77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.764735 4770 scope.go:117] "RemoveContainer" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.765102 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": container with ID starting with 6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85 not found: ID does not exist" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.765130 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} err="failed to get container status \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": rpc error: code = NotFound desc = could not find container \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": container with ID starting with 6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.765147 4770 scope.go:117] "RemoveContainer" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: E1209 11:44:39.765681 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": container with ID starting with 3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb not found: ID does not exist" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.765701 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} err="failed to get container status \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": rpc error: code = NotFound desc = could not find container \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": container with ID starting with 3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.765754 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.766091 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} err="failed to get container status \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.766107 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.766787 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} err="failed to get container status \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": rpc error: code = NotFound desc = could not find container \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": container with ID starting with 395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.766822 4770 scope.go:117] "RemoveContainer" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767122 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} err="failed to get container status \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": rpc error: code = NotFound desc = could not find container \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": container with ID starting with 79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767150 4770 scope.go:117] "RemoveContainer" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767406 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} err="failed to get container status \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": rpc error: code = NotFound desc = could not find container \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": container with ID starting with 416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767430 4770 scope.go:117] "RemoveContainer" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767660 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} err="failed to get container status \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": rpc error: code = NotFound desc = could not find container \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": container with ID starting with 37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.767683 4770 scope.go:117] "RemoveContainer" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768014 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} err="failed to get container status \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": rpc error: code = NotFound desc = could not find container \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": container with ID starting with 1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768036 4770 scope.go:117] "RemoveContainer" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768262 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} err="failed to get container status \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": rpc error: code = NotFound desc = could not find container \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": container with ID starting with 11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768279 4770 scope.go:117] "RemoveContainer" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768452 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} err="failed to get container status \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": rpc error: code = NotFound desc = could not find container \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": container with ID starting with 77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768470 4770 scope.go:117] "RemoveContainer" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768674 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} err="failed to get container status \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": rpc error: code = NotFound desc = could not find container \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": container with ID starting with 6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.768730 4770 scope.go:117] "RemoveContainer" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769061 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} err="failed to get container status \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": rpc error: code = NotFound desc = could not find container \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": container with ID starting with 3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769083 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769364 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} err="failed to get container status \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769405 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769732 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} err="failed to get container status \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": rpc error: code = NotFound desc = could not find container \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": container with ID starting with 395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.769751 4770 scope.go:117] "RemoveContainer" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770045 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} err="failed to get container status \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": rpc error: code = NotFound desc = could not find container \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": container with ID starting with 79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770066 4770 scope.go:117] "RemoveContainer" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770323 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} err="failed to get container status \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": rpc error: code = NotFound desc = could not find container \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": container with ID starting with 416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770348 4770 scope.go:117] "RemoveContainer" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770795 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} err="failed to get container status \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": rpc error: code = NotFound desc = could not find container \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": container with ID starting with 37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.770861 4770 scope.go:117] "RemoveContainer" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.771227 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} err="failed to get container status \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": rpc error: code = NotFound desc = could not find container \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": container with ID starting with 1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.771250 4770 scope.go:117] "RemoveContainer" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.771677 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} err="failed to get container status \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": rpc error: code = NotFound desc = could not find container \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": container with ID starting with 11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.771698 4770 scope.go:117] "RemoveContainer" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772045 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} err="failed to get container status \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": rpc error: code = NotFound desc = could not find container \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": container with ID starting with 77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772068 4770 scope.go:117] "RemoveContainer" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772298 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} err="failed to get container status \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": rpc error: code = NotFound desc = could not find container \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": container with ID starting with 6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772326 4770 scope.go:117] "RemoveContainer" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772580 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} err="failed to get container status \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": rpc error: code = NotFound desc = could not find container \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": container with ID starting with 3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772604 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772863 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} err="failed to get container status \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.772885 4770 scope.go:117] "RemoveContainer" containerID="395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773163 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b"} err="failed to get container status \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": rpc error: code = NotFound desc = could not find container \"395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b\": container with ID starting with 395959a09b316b168507dc66e173dc9eee02db1d9985f4ea96783b3a5add8d0b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773183 4770 scope.go:117] "RemoveContainer" containerID="79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773455 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41"} err="failed to get container status \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": rpc error: code = NotFound desc = could not find container \"79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41\": container with ID starting with 79573f80115a8d0bb374f74ca164b501b5f28d8299767085beaf3fb69f2e8b41 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773473 4770 scope.go:117] "RemoveContainer" containerID="416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773701 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8"} err="failed to get container status \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": rpc error: code = NotFound desc = could not find container \"416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8\": container with ID starting with 416c20ef95039c171ff69ce4fa790fe71ac87d3ddeb635b0a478ff51d2bebac8 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773722 4770 scope.go:117] "RemoveContainer" containerID="37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773964 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6"} err="failed to get container status \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": rpc error: code = NotFound desc = could not find container \"37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6\": container with ID starting with 37af4d43385cc9bf31dff33f580062b104c0cc78e647a2f6e8d79988c1ecdda6 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.773984 4770 scope.go:117] "RemoveContainer" containerID="1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774226 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e"} err="failed to get container status \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": rpc error: code = NotFound desc = could not find container \"1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e\": container with ID starting with 1adc661fe52eadc7164ae2f5aafc4921a593f0e45b3da97ec1b1684d315dee3e not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774243 4770 scope.go:117] "RemoveContainer" containerID="11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774496 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b"} err="failed to get container status \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": rpc error: code = NotFound desc = could not find container \"11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b\": container with ID starting with 11253a5e13ac37b9d76f43ea9c78fbf9ed45f04f5816ab0d5c4459aec0e30e5b not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774516 4770 scope.go:117] "RemoveContainer" containerID="77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774784 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5"} err="failed to get container status \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": rpc error: code = NotFound desc = could not find container \"77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5\": container with ID starting with 77219b4f9d9d2f110fd8ba464c6c120c3a8de187b6d99d11bf8dfa0c2b1e60a5 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.774801 4770 scope.go:117] "RemoveContainer" containerID="6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.775096 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85"} err="failed to get container status \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": rpc error: code = NotFound desc = could not find container \"6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85\": container with ID starting with 6e5700060cce0e633ca3e1215c7ebd9c4da5c6d2be478b5b588ec9794d175f85 not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.775156 4770 scope.go:117] "RemoveContainer" containerID="3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.775517 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb"} err="failed to get container status \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": rpc error: code = NotFound desc = could not find container \"3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb\": container with ID starting with 3b394a35053343399620164335e8c037f3d0497da53d732b19d7b3a92652c8eb not found: ID does not exist" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.775538 4770 scope.go:117] "RemoveContainer" containerID="33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865" Dec 09 11:44:39 crc kubenswrapper[4770]: I1209 11:44:39.775746 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865"} err="failed to get container status \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": rpc error: code = NotFound desc = could not find container \"33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865\": container with ID starting with 33567c3a74cb43bf4a27473f7df4e689746e2152e8270821578c2d279fa8b865 not found: ID does not exist" Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"d6d04f6833a5a7b3814a33e9d1b9e8215f55a8ac308fa0b0a87e911281a4babd"} Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479434 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"b219b65027d532f731e406927b00b2ec60cd5845db4495784c64f890ded57334"} Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479450 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"fb57388623d8a3080366de5539e53e8325887b6bbac75e534c08f00c925aaad1"} Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479460 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"6416e9f3c63464efb0b9ff5e2618fbe392a15e494f79d11c6872d0588ebfd682"} Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"b1e76ba8c22d1e85207e5a72dca44df98964b94308c70ee5ec9c32f56e5b2cd2"} Dec 09 11:44:40 crc kubenswrapper[4770]: I1209 11:44:40.479485 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"f2de1466c6207a47d336f3ed576dcc9917bce0b32e488a970cfc2da58ce9bb43"} Dec 09 11:44:41 crc kubenswrapper[4770]: I1209 11:44:41.348649 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa774d4-1a7d-4731-a47b-d1d97c88869e" path="/var/lib/kubelet/pods/eaa774d4-1a7d-4731-a47b-d1d97c88869e/volumes" Dec 09 11:44:43 crc kubenswrapper[4770]: I1209 11:44:43.501603 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"082264a690745dde3926ad83e42fcf274c02bdd9bd5701398fcfc5d0fd47031d"} Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.710066 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-5rknq"] Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.711417 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.715798 4770 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-rlzzh" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.716018 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.716235 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.717259 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.808711 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.808809 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6s9\" (UniqueName: \"kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.808862 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.910591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.910695 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6s9\" (UniqueName: \"kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.910742 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.911091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.912396 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:45 crc kubenswrapper[4770]: I1209 11:44:45.936716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6s9\" (UniqueName: \"kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9\") pod \"crc-storage-crc-5rknq\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:46 crc kubenswrapper[4770]: I1209 11:44:46.040634 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:46 crc kubenswrapper[4770]: E1209 11:44:46.078586 4770 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(09528aa160a2bc474cc9d6f26749e67c432cc1b9029bbe8076f24ec67169894b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:44:46 crc kubenswrapper[4770]: E1209 11:44:46.078727 4770 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(09528aa160a2bc474cc9d6f26749e67c432cc1b9029bbe8076f24ec67169894b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:46 crc kubenswrapper[4770]: E1209 11:44:46.078760 4770 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(09528aa160a2bc474cc9d6f26749e67c432cc1b9029bbe8076f24ec67169894b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:46 crc kubenswrapper[4770]: E1209 11:44:46.078812 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5rknq_crc-storage(48936f2c-67ec-4a4c-a08d-3713724df885)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5rknq_crc-storage(48936f2c-67ec-4a4c-a08d-3713724df885)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(09528aa160a2bc474cc9d6f26749e67c432cc1b9029bbe8076f24ec67169894b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5rknq" podUID="48936f2c-67ec-4a4c-a08d-3713724df885" Dec 09 11:44:46 crc kubenswrapper[4770]: I1209 11:44:46.519616 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" event={"ID":"ceb5b168-5d89-413c-b590-0b29d1a60b29","Type":"ContainerStarted","Data":"6d63e0b04eaf8d57022068a56de011e995ab43d6b5d1899fdcb6b19d10b75b1f"} Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.529929 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.530534 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.530545 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.562310 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.563182 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" podStartSLOduration=10.563156216 podStartE2EDuration="10.563156216s" podCreationTimestamp="2025-12-09 11:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:44:48.559212856 +0000 UTC m=+813.799971375" watchObservedRunningTime="2025-12-09 11:44:48.563156216 +0000 UTC m=+813.803914735" Dec 09 11:44:48 crc kubenswrapper[4770]: I1209 11:44:48.566649 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:44:49 crc kubenswrapper[4770]: I1209 11:44:49.230360 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5rknq"] Dec 09 11:44:49 crc kubenswrapper[4770]: I1209 11:44:49.231111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:49 crc kubenswrapper[4770]: I1209 11:44:49.231760 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:49 crc kubenswrapper[4770]: E1209 11:44:49.263815 4770 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(a42a44ee0943ee43dd104d36c957ff53357391715fec9ef26a628260c42cb75a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 11:44:49 crc kubenswrapper[4770]: E1209 11:44:49.263888 4770 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(a42a44ee0943ee43dd104d36c957ff53357391715fec9ef26a628260c42cb75a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:49 crc kubenswrapper[4770]: E1209 11:44:49.263928 4770 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(a42a44ee0943ee43dd104d36c957ff53357391715fec9ef26a628260c42cb75a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:44:49 crc kubenswrapper[4770]: E1209 11:44:49.264045 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5rknq_crc-storage(48936f2c-67ec-4a4c-a08d-3713724df885)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5rknq_crc-storage(48936f2c-67ec-4a4c-a08d-3713724df885)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5rknq_crc-storage_48936f2c-67ec-4a4c-a08d-3713724df885_0(a42a44ee0943ee43dd104d36c957ff53357391715fec9ef26a628260c42cb75a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5rknq" podUID="48936f2c-67ec-4a4c-a08d-3713724df885" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.166535 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk"] Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.168618 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.171243 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.172332 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.185198 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk"] Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.225489 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.225645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kph5c\" (UniqueName: \"kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.225695 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.327516 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kph5c\" (UniqueName: \"kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.327599 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.327635 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.328796 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.338627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.346742 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kph5c\" (UniqueName: \"kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c\") pod \"collect-profiles-29421345-7stsk\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.501388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:00 crc kubenswrapper[4770]: I1209 11:45:00.745015 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk"] Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.340173 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.340991 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.563422 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5rknq"] Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.575410 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.626756 4770 generic.go:334] "Generic (PLEG): container finished" podID="b81a4879-e6d0-4eb8-b823-93b45c1fb249" containerID="53d9de51c2168967fc90a6f74c3d89222c199ce1f88cd29a368f94acdaf27178" exitCode=0 Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.626877 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" event={"ID":"b81a4879-e6d0-4eb8-b823-93b45c1fb249","Type":"ContainerDied","Data":"53d9de51c2168967fc90a6f74c3d89222c199ce1f88cd29a368f94acdaf27178"} Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.626983 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" event={"ID":"b81a4879-e6d0-4eb8-b823-93b45c1fb249","Type":"ContainerStarted","Data":"aabbd24e008839688483dc640c5a1d6a45e20335d6b6d854629141687ddd8622"} Dec 09 11:45:01 crc kubenswrapper[4770]: I1209 11:45:01.628390 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5rknq" event={"ID":"48936f2c-67ec-4a4c-a08d-3713724df885","Type":"ContainerStarted","Data":"8bbe1b7b9cc8dfd02452d28374ef6268559c404590e2400764acf7884b5be43b"} Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.877838 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.975505 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume\") pod \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.975583 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume\") pod \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.975651 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kph5c\" (UniqueName: \"kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c\") pod \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\" (UID: \"b81a4879-e6d0-4eb8-b823-93b45c1fb249\") " Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.976988 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume" (OuterVolumeSpecName: "config-volume") pod "b81a4879-e6d0-4eb8-b823-93b45c1fb249" (UID: "b81a4879-e6d0-4eb8-b823-93b45c1fb249"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.981540 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c" (OuterVolumeSpecName: "kube-api-access-kph5c") pod "b81a4879-e6d0-4eb8-b823-93b45c1fb249" (UID: "b81a4879-e6d0-4eb8-b823-93b45c1fb249"). InnerVolumeSpecName "kube-api-access-kph5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:45:02 crc kubenswrapper[4770]: I1209 11:45:02.981785 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b81a4879-e6d0-4eb8-b823-93b45c1fb249" (UID: "b81a4879-e6d0-4eb8-b823-93b45c1fb249"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.077171 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b81a4879-e6d0-4eb8-b823-93b45c1fb249-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.077662 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b81a4879-e6d0-4eb8-b823-93b45c1fb249-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.077678 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kph5c\" (UniqueName: \"kubernetes.io/projected/b81a4879-e6d0-4eb8-b823-93b45c1fb249-kube-api-access-kph5c\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.644698 4770 generic.go:334] "Generic (PLEG): container finished" podID="48936f2c-67ec-4a4c-a08d-3713724df885" containerID="280fc29c8db2a076cc81cc1692b6192d89f2ac26d27873d9b0596a62fc959422" exitCode=0 Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.644764 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5rknq" event={"ID":"48936f2c-67ec-4a4c-a08d-3713724df885","Type":"ContainerDied","Data":"280fc29c8db2a076cc81cc1692b6192d89f2ac26d27873d9b0596a62fc959422"} Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.651199 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" event={"ID":"b81a4879-e6d0-4eb8-b823-93b45c1fb249","Type":"ContainerDied","Data":"aabbd24e008839688483dc640c5a1d6a45e20335d6b6d854629141687ddd8622"} Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.651256 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aabbd24e008839688483dc640c5a1d6a45e20335d6b6d854629141687ddd8622" Dec 09 11:45:03 crc kubenswrapper[4770]: I1209 11:45:03.651359 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk" Dec 09 11:45:04 crc kubenswrapper[4770]: I1209 11:45:04.889891 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.005365 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt\") pod \"48936f2c-67ec-4a4c-a08d-3713724df885\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.005870 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "48936f2c-67ec-4a4c-a08d-3713724df885" (UID: "48936f2c-67ec-4a4c-a08d-3713724df885"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.006148 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c6s9\" (UniqueName: \"kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9\") pod \"48936f2c-67ec-4a4c-a08d-3713724df885\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.006179 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage\") pod \"48936f2c-67ec-4a4c-a08d-3713724df885\" (UID: \"48936f2c-67ec-4a4c-a08d-3713724df885\") " Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.007084 4770 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/48936f2c-67ec-4a4c-a08d-3713724df885-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.025998 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9" (OuterVolumeSpecName: "kube-api-access-7c6s9") pod "48936f2c-67ec-4a4c-a08d-3713724df885" (UID: "48936f2c-67ec-4a4c-a08d-3713724df885"). InnerVolumeSpecName "kube-api-access-7c6s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.028211 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "48936f2c-67ec-4a4c-a08d-3713724df885" (UID: "48936f2c-67ec-4a4c-a08d-3713724df885"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.108884 4770 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/48936f2c-67ec-4a4c-a08d-3713724df885-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.108942 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c6s9\" (UniqueName: \"kubernetes.io/projected/48936f2c-67ec-4a4c-a08d-3713724df885-kube-api-access-7c6s9\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.662692 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5rknq" event={"ID":"48936f2c-67ec-4a4c-a08d-3713724df885","Type":"ContainerDied","Data":"8bbe1b7b9cc8dfd02452d28374ef6268559c404590e2400764acf7884b5be43b"} Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.662753 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bbe1b7b9cc8dfd02452d28374ef6268559c404590e2400764acf7884b5be43b" Dec 09 11:45:05 crc kubenswrapper[4770]: I1209 11:45:05.662776 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5rknq" Dec 09 11:45:09 crc kubenswrapper[4770]: I1209 11:45:09.097492 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nshst" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.500874 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2"] Dec 09 11:45:12 crc kubenswrapper[4770]: E1209 11:45:12.501512 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48936f2c-67ec-4a4c-a08d-3713724df885" containerName="storage" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.501528 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="48936f2c-67ec-4a4c-a08d-3713724df885" containerName="storage" Dec 09 11:45:12 crc kubenswrapper[4770]: E1209 11:45:12.501544 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81a4879-e6d0-4eb8-b823-93b45c1fb249" containerName="collect-profiles" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.501552 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81a4879-e6d0-4eb8-b823-93b45c1fb249" containerName="collect-profiles" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.501679 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="48936f2c-67ec-4a4c-a08d-3713724df885" containerName="storage" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.501698 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b81a4879-e6d0-4eb8-b823-93b45c1fb249" containerName="collect-profiles" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.502509 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.504953 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.524429 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2"] Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.622001 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.622124 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.622341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzg5\" (UniqueName: \"kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.724432 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.724548 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzg5\" (UniqueName: \"kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.724605 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.725064 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.725096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.751588 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzg5\" (UniqueName: \"kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:12 crc kubenswrapper[4770]: I1209 11:45:12.820550 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:13 crc kubenswrapper[4770]: I1209 11:45:13.041375 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2"] Dec 09 11:45:13 crc kubenswrapper[4770]: I1209 11:45:13.706356 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerStarted","Data":"82b2b92fc67f6d2f021b3bd23ab6bf8eacb5b0bdcf1e0dd5a5cb6daf78940d80"} Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.196589 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.203986 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.205131 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.351569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.351691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hng8x\" (UniqueName: \"kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.351753 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.453149 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hng8x\" (UniqueName: \"kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.453671 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.453718 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.454673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.454737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.475852 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hng8x\" (UniqueName: \"kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x\") pod \"redhat-operators-7qrkz\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.519859 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.718496 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerStarted","Data":"66b8f03ee6db957156e5f00f6a0cfc4bb5e7b7061c1cae8aa04c4d376b4ad301"} Dec 09 11:45:14 crc kubenswrapper[4770]: I1209 11:45:14.766471 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:15 crc kubenswrapper[4770]: I1209 11:45:15.725652 4770 generic.go:334] "Generic (PLEG): container finished" podID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerID="66b8f03ee6db957156e5f00f6a0cfc4bb5e7b7061c1cae8aa04c4d376b4ad301" exitCode=0 Dec 09 11:45:15 crc kubenswrapper[4770]: I1209 11:45:15.725708 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerDied","Data":"66b8f03ee6db957156e5f00f6a0cfc4bb5e7b7061c1cae8aa04c4d376b4ad301"} Dec 09 11:45:15 crc kubenswrapper[4770]: I1209 11:45:15.728170 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerID="6e706c9c6eeaa4fc50b5e597429bd294e3d72ab814ad7185c0945ff972429259" exitCode=0 Dec 09 11:45:15 crc kubenswrapper[4770]: I1209 11:45:15.728206 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerDied","Data":"6e706c9c6eeaa4fc50b5e597429bd294e3d72ab814ad7185c0945ff972429259"} Dec 09 11:45:15 crc kubenswrapper[4770]: I1209 11:45:15.728233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerStarted","Data":"7e72e2bfd0bb2092d827e2ed4d64f2bf6b85bdf8bf99ef4d265bd8711f942e62"} Dec 09 11:45:16 crc kubenswrapper[4770]: I1209 11:45:16.738941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerStarted","Data":"ed1d3817d3b9ab62569e7bd872cb26dba0e32242092732ee59e6630e73623045"} Dec 09 11:45:17 crc kubenswrapper[4770]: I1209 11:45:17.748135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerStarted","Data":"be571807536d4e8478435ac2d4175b161efe233eebcc18b68d9f8ef3354565f3"} Dec 09 11:45:17 crc kubenswrapper[4770]: I1209 11:45:17.750873 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerID="ed1d3817d3b9ab62569e7bd872cb26dba0e32242092732ee59e6630e73623045" exitCode=0 Dec 09 11:45:17 crc kubenswrapper[4770]: I1209 11:45:17.750966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerDied","Data":"ed1d3817d3b9ab62569e7bd872cb26dba0e32242092732ee59e6630e73623045"} Dec 09 11:45:18 crc kubenswrapper[4770]: I1209 11:45:18.766305 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerStarted","Data":"28e7f73bf411e31bf9e64ba719fcb7c982b79f9a4f161a766588c1dc83d0411f"} Dec 09 11:45:18 crc kubenswrapper[4770]: I1209 11:45:18.768543 4770 generic.go:334] "Generic (PLEG): container finished" podID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerID="be571807536d4e8478435ac2d4175b161efe233eebcc18b68d9f8ef3354565f3" exitCode=0 Dec 09 11:45:18 crc kubenswrapper[4770]: I1209 11:45:18.768580 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerDied","Data":"be571807536d4e8478435ac2d4175b161efe233eebcc18b68d9f8ef3354565f3"} Dec 09 11:45:18 crc kubenswrapper[4770]: I1209 11:45:18.793643 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7qrkz" podStartSLOduration=2.318007625 podStartE2EDuration="4.793617979s" podCreationTimestamp="2025-12-09 11:45:14 +0000 UTC" firstStartedPulling="2025-12-09 11:45:15.729858423 +0000 UTC m=+840.970616942" lastFinishedPulling="2025-12-09 11:45:18.205468777 +0000 UTC m=+843.446227296" observedRunningTime="2025-12-09 11:45:18.78655837 +0000 UTC m=+844.027316889" watchObservedRunningTime="2025-12-09 11:45:18.793617979 +0000 UTC m=+844.034376498" Dec 09 11:45:19 crc kubenswrapper[4770]: I1209 11:45:19.778032 4770 generic.go:334] "Generic (PLEG): container finished" podID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerID="e0cc181162c6a70a60cbcac6cacbb6d51456890e09bff34a67c099673806a43f" exitCode=0 Dec 09 11:45:19 crc kubenswrapper[4770]: I1209 11:45:19.778328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerDied","Data":"e0cc181162c6a70a60cbcac6cacbb6d51456890e09bff34a67c099673806a43f"} Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.052609 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.156509 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kzg5\" (UniqueName: \"kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5\") pod \"91ff00b8-1682-47e7-99d8-b0444921b1a1\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.156611 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util\") pod \"91ff00b8-1682-47e7-99d8-b0444921b1a1\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.156793 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle\") pod \"91ff00b8-1682-47e7-99d8-b0444921b1a1\" (UID: \"91ff00b8-1682-47e7-99d8-b0444921b1a1\") " Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.157848 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle" (OuterVolumeSpecName: "bundle") pod "91ff00b8-1682-47e7-99d8-b0444921b1a1" (UID: "91ff00b8-1682-47e7-99d8-b0444921b1a1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.162949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5" (OuterVolumeSpecName: "kube-api-access-8kzg5") pod "91ff00b8-1682-47e7-99d8-b0444921b1a1" (UID: "91ff00b8-1682-47e7-99d8-b0444921b1a1"). InnerVolumeSpecName "kube-api-access-8kzg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.168568 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util" (OuterVolumeSpecName: "util") pod "91ff00b8-1682-47e7-99d8-b0444921b1a1" (UID: "91ff00b8-1682-47e7-99d8-b0444921b1a1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.258800 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.258850 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kzg5\" (UniqueName: \"kubernetes.io/projected/91ff00b8-1682-47e7-99d8-b0444921b1a1-kube-api-access-8kzg5\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.258868 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/91ff00b8-1682-47e7-99d8-b0444921b1a1-util\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.795793 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" event={"ID":"91ff00b8-1682-47e7-99d8-b0444921b1a1","Type":"ContainerDied","Data":"82b2b92fc67f6d2f021b3bd23ab6bf8eacb5b0bdcf1e0dd5a5cb6daf78940d80"} Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.795863 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82b2b92fc67f6d2f021b3bd23ab6bf8eacb5b0bdcf1e0dd5a5cb6daf78940d80" Dec 09 11:45:21 crc kubenswrapper[4770]: I1209 11:45:21.795877 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.511275 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf"] Dec 09 11:45:23 crc kubenswrapper[4770]: E1209 11:45:23.512274 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="pull" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.512369 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="pull" Dec 09 11:45:23 crc kubenswrapper[4770]: E1209 11:45:23.512443 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="extract" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.512526 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="extract" Dec 09 11:45:23 crc kubenswrapper[4770]: E1209 11:45:23.512592 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="util" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.512643 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="util" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.512798 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ff00b8-1682-47e7-99d8-b0444921b1a1" containerName="extract" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.513335 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.516022 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-n6fv9" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.516450 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.517446 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.527206 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf"] Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.588686 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5fl\" (UniqueName: \"kubernetes.io/projected/094c86c2-b555-44b7-b26f-6915f0a5eaa5-kube-api-access-rt5fl\") pod \"nmstate-operator-5b5b58f5c8-twvbf\" (UID: \"094c86c2-b555-44b7-b26f-6915f0a5eaa5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.689840 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5fl\" (UniqueName: \"kubernetes.io/projected/094c86c2-b555-44b7-b26f-6915f0a5eaa5-kube-api-access-rt5fl\") pod \"nmstate-operator-5b5b58f5c8-twvbf\" (UID: \"094c86c2-b555-44b7-b26f-6915f0a5eaa5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.712684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5fl\" (UniqueName: \"kubernetes.io/projected/094c86c2-b555-44b7-b26f-6915f0a5eaa5-kube-api-access-rt5fl\") pod \"nmstate-operator-5b5b58f5c8-twvbf\" (UID: \"094c86c2-b555-44b7-b26f-6915f0a5eaa5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" Dec 09 11:45:23 crc kubenswrapper[4770]: I1209 11:45:23.833744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.194190 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf"] Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.520770 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.521204 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.579354 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.816343 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" event={"ID":"094c86c2-b555-44b7-b26f-6915f0a5eaa5","Type":"ContainerStarted","Data":"a6f1069a28adf7a4e1752f39298efc69562d0650f66c78168b5f4c8b63d08797"} Dec 09 11:45:24 crc kubenswrapper[4770]: I1209 11:45:24.865978 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:26 crc kubenswrapper[4770]: I1209 11:45:26.980071 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:26 crc kubenswrapper[4770]: I1209 11:45:26.980699 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7qrkz" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="registry-server" containerID="cri-o://28e7f73bf411e31bf9e64ba719fcb7c982b79f9a4f161a766588c1dc83d0411f" gracePeriod=2 Dec 09 11:45:29 crc kubenswrapper[4770]: I1209 11:45:29.876315 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerID="28e7f73bf411e31bf9e64ba719fcb7c982b79f9a4f161a766588c1dc83d0411f" exitCode=0 Dec 09 11:45:29 crc kubenswrapper[4770]: I1209 11:45:29.876719 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerDied","Data":"28e7f73bf411e31bf9e64ba719fcb7c982b79f9a4f161a766588c1dc83d0411f"} Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.061237 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.234541 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities\") pod \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.234665 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content\") pod \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.234700 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hng8x\" (UniqueName: \"kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x\") pod \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\" (UID: \"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1\") " Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.235839 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities" (OuterVolumeSpecName: "utilities") pod "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" (UID: "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.249619 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x" (OuterVolumeSpecName: "kube-api-access-hng8x") pod "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" (UID: "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1"). InnerVolumeSpecName "kube-api-access-hng8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.336499 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hng8x\" (UniqueName: \"kubernetes.io/projected/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-kube-api-access-hng8x\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.336556 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.349821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" (UID: "8c0d386e-9764-4e29-8eb6-505e8f4bdfd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.438807 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.885351 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" event={"ID":"094c86c2-b555-44b7-b26f-6915f0a5eaa5","Type":"ContainerStarted","Data":"5d37857ba8e5681f9c24ac5c88e38b0b133459ca5ac228e68fe6cfa58e8850ed"} Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.888169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qrkz" event={"ID":"8c0d386e-9764-4e29-8eb6-505e8f4bdfd1","Type":"ContainerDied","Data":"7e72e2bfd0bb2092d827e2ed4d64f2bf6b85bdf8bf99ef4d265bd8711f942e62"} Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.888215 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qrkz" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.888267 4770 scope.go:117] "RemoveContainer" containerID="28e7f73bf411e31bf9e64ba719fcb7c982b79f9a4f161a766588c1dc83d0411f" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.908202 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-twvbf" podStartSLOduration=2.323469355 podStartE2EDuration="7.908174788s" podCreationTimestamp="2025-12-09 11:45:23 +0000 UTC" firstStartedPulling="2025-12-09 11:45:24.21331633 +0000 UTC m=+849.454074859" lastFinishedPulling="2025-12-09 11:45:29.798021773 +0000 UTC m=+855.038780292" observedRunningTime="2025-12-09 11:45:30.905694635 +0000 UTC m=+856.146453164" watchObservedRunningTime="2025-12-09 11:45:30.908174788 +0000 UTC m=+856.148933317" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.916606 4770 scope.go:117] "RemoveContainer" containerID="ed1d3817d3b9ab62569e7bd872cb26dba0e32242092732ee59e6630e73623045" Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.923086 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.929588 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7qrkz"] Dec 09 11:45:30 crc kubenswrapper[4770]: I1209 11:45:30.942962 4770 scope.go:117] "RemoveContainer" containerID="6e706c9c6eeaa4fc50b5e597429bd294e3d72ab814ad7185c0945ff972429259" Dec 09 11:45:31 crc kubenswrapper[4770]: I1209 11:45:31.348639 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" path="/var/lib/kubelet/pods/8c0d386e-9764-4e29-8eb6-505e8f4bdfd1/volumes" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.587463 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs"] Dec 09 11:45:32 crc kubenswrapper[4770]: E1209 11:45:32.587870 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="extract-utilities" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.587888 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="extract-utilities" Dec 09 11:45:32 crc kubenswrapper[4770]: E1209 11:45:32.587950 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="extract-content" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.587963 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="extract-content" Dec 09 11:45:32 crc kubenswrapper[4770]: E1209 11:45:32.587975 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="registry-server" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.587985 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="registry-server" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.588114 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0d386e-9764-4e29-8eb6-505e8f4bdfd1" containerName="registry-server" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.589018 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.592016 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.592553 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-st8xd" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.593197 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.596970 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.612644 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.617020 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.634068 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-dpkk8"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.634999 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.669508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjhwp\" (UniqueName: \"kubernetes.io/projected/86ad5bcc-54e9-4125-a7d7-db0c74de1b01-kube-api-access-vjhwp\") pod \"nmstate-metrics-7f946cbc9-8s7rs\" (UID: \"86ad5bcc-54e9-4125-a7d7-db0c74de1b01\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.669622 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ab4bae62-c176-4ff2-89af-2b9a362c4f65-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.669657 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvvhb\" (UniqueName: \"kubernetes.io/projected/ab4bae62-c176-4ff2-89af-2b9a362c4f65-kube-api-access-jvvhb\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.783940 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786137 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786525 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ab4bae62-c176-4ff2-89af-2b9a362c4f65-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786594 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-ovs-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786635 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvvhb\" (UniqueName: \"kubernetes.io/projected/ab4bae62-c176-4ff2-89af-2b9a362c4f65-kube-api-access-jvvhb\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-nmstate-lock\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.786857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vjmd\" (UniqueName: \"kubernetes.io/projected/e48df172-3e28-47a9-acf7-80d69988cd6b-kube-api-access-8vjmd\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.787030 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjhwp\" (UniqueName: \"kubernetes.io/projected/86ad5bcc-54e9-4125-a7d7-db0c74de1b01-kube-api-access-vjhwp\") pod \"nmstate-metrics-7f946cbc9-8s7rs\" (UID: \"86ad5bcc-54e9-4125-a7d7-db0c74de1b01\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.787062 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-dbus-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.790450 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.790958 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-k2shq" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.791127 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.806762 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjhwp\" (UniqueName: \"kubernetes.io/projected/86ad5bcc-54e9-4125-a7d7-db0c74de1b01-kube-api-access-vjhwp\") pod \"nmstate-metrics-7f946cbc9-8s7rs\" (UID: \"86ad5bcc-54e9-4125-a7d7-db0c74de1b01\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.812858 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ab4bae62-c176-4ff2-89af-2b9a362c4f65-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.814103 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvvhb\" (UniqueName: \"kubernetes.io/projected/ab4bae62-c176-4ff2-89af-2b9a362c4f65-kube-api-access-jvvhb\") pod \"nmstate-webhook-5f6d4c5ccb-6p6bz\" (UID: \"ab4bae62-c176-4ff2-89af-2b9a362c4f65\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.820192 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv"] Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888413 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-ovs-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e684363b-6fd1-4432-8cee-cdb9858e4367-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888569 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-nmstate-lock\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888593 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spcbc\" (UniqueName: \"kubernetes.io/projected/e684363b-6fd1-4432-8cee-cdb9858e4367-kube-api-access-spcbc\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888629 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vjmd\" (UniqueName: \"kubernetes.io/projected/e48df172-3e28-47a9-acf7-80d69988cd6b-kube-api-access-8vjmd\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.888725 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-dbus-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.889301 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-dbus-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.889390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-ovs-socket\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.889440 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e48df172-3e28-47a9-acf7-80d69988cd6b-nmstate-lock\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.913802 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.920335 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vjmd\" (UniqueName: \"kubernetes.io/projected/e48df172-3e28-47a9-acf7-80d69988cd6b-kube-api-access-8vjmd\") pod \"nmstate-handler-dpkk8\" (UID: \"e48df172-3e28-47a9-acf7-80d69988cd6b\") " pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.923883 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.964301 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.990292 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e684363b-6fd1-4432-8cee-cdb9858e4367-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.990364 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spcbc\" (UniqueName: \"kubernetes.io/projected/e684363b-6fd1-4432-8cee-cdb9858e4367-kube-api-access-spcbc\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.990399 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:32 crc kubenswrapper[4770]: E1209 11:45:32.990579 4770 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 09 11:45:32 crc kubenswrapper[4770]: E1209 11:45:32.990666 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert podName:e684363b-6fd1-4432-8cee-cdb9858e4367 nodeName:}" failed. No retries permitted until 2025-12-09 11:45:33.490630339 +0000 UTC m=+858.731388858 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-84prv" (UID: "e684363b-6fd1-4432-8cee-cdb9858e4367") : secret "plugin-serving-cert" not found Dec 09 11:45:32 crc kubenswrapper[4770]: I1209 11:45:32.992104 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e684363b-6fd1-4432-8cee-cdb9858e4367-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.018198 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-68dc874465-hzpvm"] Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.022930 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.034272 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68dc874465-hzpvm"] Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.050664 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spcbc\" (UniqueName: \"kubernetes.io/projected/e684363b-6fd1-4432-8cee-cdb9858e4367-kube-api-access-spcbc\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196018 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-oauth-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196082 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-trusted-ca-bundle\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196117 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-oauth-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196145 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196441 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-service-ca\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196555 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-console-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.196590 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8wfl\" (UniqueName: \"kubernetes.io/projected/aa03cc21-3379-4726-b035-f33dcc6799d9-kube-api-access-t8wfl\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.239049 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs"] Dec 09 11:45:33 crc kubenswrapper[4770]: W1209 11:45:33.249429 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86ad5bcc_54e9_4125_a7d7_db0c74de1b01.slice/crio-c5dfbce0d467e22bc2dd65fcc23b449a8549a29f38d10d1b32c6a03069278f6c WatchSource:0}: Error finding container c5dfbce0d467e22bc2dd65fcc23b449a8549a29f38d10d1b32c6a03069278f6c: Status 404 returned error can't find the container with id c5dfbce0d467e22bc2dd65fcc23b449a8549a29f38d10d1b32c6a03069278f6c Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.292036 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz"] Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298597 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-service-ca\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298657 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-console-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8wfl\" (UniqueName: \"kubernetes.io/projected/aa03cc21-3379-4726-b035-f33dcc6799d9-kube-api-access-t8wfl\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-trusted-ca-bundle\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298823 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-oauth-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298850 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-oauth-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.298957 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.300978 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-console-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.301077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-service-ca\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.301121 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-oauth-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.301172 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa03cc21-3379-4726-b035-f33dcc6799d9-trusted-ca-bundle\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: W1209 11:45:33.302276 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab4bae62_c176_4ff2_89af_2b9a362c4f65.slice/crio-b25b3c50c71b6bd3da9a77f182953523ed1ee8f60283c62affd4a197b16ffede WatchSource:0}: Error finding container b25b3c50c71b6bd3da9a77f182953523ed1ee8f60283c62affd4a197b16ffede: Status 404 returned error can't find the container with id b25b3c50c71b6bd3da9a77f182953523ed1ee8f60283c62affd4a197b16ffede Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.304706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-serving-cert\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.310060 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aa03cc21-3379-4726-b035-f33dcc6799d9-console-oauth-config\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.321649 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8wfl\" (UniqueName: \"kubernetes.io/projected/aa03cc21-3379-4726-b035-f33dcc6799d9-kube-api-access-t8wfl\") pod \"console-68dc874465-hzpvm\" (UID: \"aa03cc21-3379-4726-b035-f33dcc6799d9\") " pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.353421 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.502061 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:33 crc kubenswrapper[4770]: E1209 11:45:33.502358 4770 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 09 11:45:33 crc kubenswrapper[4770]: E1209 11:45:33.502431 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert podName:e684363b-6fd1-4432-8cee-cdb9858e4367 nodeName:}" failed. No retries permitted until 2025-12-09 11:45:34.502409887 +0000 UTC m=+859.743168406 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-84prv" (UID: "e684363b-6fd1-4432-8cee-cdb9858e4367") : secret "plugin-serving-cert" not found Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.678100 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68dc874465-hzpvm"] Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.919136 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dpkk8" event={"ID":"e48df172-3e28-47a9-acf7-80d69988cd6b","Type":"ContainerStarted","Data":"f9e54be1ff0fa271efeed81f7fd00074565053bc15c39ed2c8a32cd0dd30ff9b"} Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.921031 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" event={"ID":"ab4bae62-c176-4ff2-89af-2b9a362c4f65","Type":"ContainerStarted","Data":"b25b3c50c71b6bd3da9a77f182953523ed1ee8f60283c62affd4a197b16ffede"} Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.922251 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" event={"ID":"86ad5bcc-54e9-4125-a7d7-db0c74de1b01","Type":"ContainerStarted","Data":"c5dfbce0d467e22bc2dd65fcc23b449a8549a29f38d10d1b32c6a03069278f6c"} Dec 09 11:45:33 crc kubenswrapper[4770]: I1209 11:45:33.923419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68dc874465-hzpvm" event={"ID":"aa03cc21-3379-4726-b035-f33dcc6799d9","Type":"ContainerStarted","Data":"54b2f46e39f8e432c09b5206750287882e3e4fca9907ce7762c872d63e51822f"} Dec 09 11:45:34 crc kubenswrapper[4770]: I1209 11:45:34.514831 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:34 crc kubenswrapper[4770]: I1209 11:45:34.535826 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e684363b-6fd1-4432-8cee-cdb9858e4367-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-84prv\" (UID: \"e684363b-6fd1-4432-8cee-cdb9858e4367\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:34 crc kubenswrapper[4770]: I1209 11:45:34.653967 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" Dec 09 11:45:34 crc kubenswrapper[4770]: I1209 11:45:34.942780 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68dc874465-hzpvm" event={"ID":"aa03cc21-3379-4726-b035-f33dcc6799d9","Type":"ContainerStarted","Data":"dd011ffe19ce3c6e8512cc8128be45e26610eed3f8a7303d93503ef90cfec0c9"} Dec 09 11:45:34 crc kubenswrapper[4770]: I1209 11:45:34.967755 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-68dc874465-hzpvm" podStartSLOduration=2.967710195 podStartE2EDuration="2.967710195s" podCreationTimestamp="2025-12-09 11:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:45:34.966124005 +0000 UTC m=+860.206882524" watchObservedRunningTime="2025-12-09 11:45:34.967710195 +0000 UTC m=+860.208468724" Dec 09 11:45:35 crc kubenswrapper[4770]: I1209 11:45:35.002939 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv"] Dec 09 11:45:35 crc kubenswrapper[4770]: W1209 11:45:35.015829 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode684363b_6fd1_4432_8cee_cdb9858e4367.slice/crio-af5eef57ef9e201bf2506d121cd219850b7824c073f20257246db109cbc6c846 WatchSource:0}: Error finding container af5eef57ef9e201bf2506d121cd219850b7824c073f20257246db109cbc6c846: Status 404 returned error can't find the container with id af5eef57ef9e201bf2506d121cd219850b7824c073f20257246db109cbc6c846 Dec 09 11:45:35 crc kubenswrapper[4770]: I1209 11:45:35.955231 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" event={"ID":"e684363b-6fd1-4432-8cee-cdb9858e4367","Type":"ContainerStarted","Data":"af5eef57ef9e201bf2506d121cd219850b7824c073f20257246db109cbc6c846"} Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.972436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dpkk8" event={"ID":"e48df172-3e28-47a9-acf7-80d69988cd6b","Type":"ContainerStarted","Data":"03e5c2b5426abee60bb711ef61e24ee9e546a91336fd761e66dea7ba1545e385"} Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.973251 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.979017 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" event={"ID":"86ad5bcc-54e9-4125-a7d7-db0c74de1b01","Type":"ContainerStarted","Data":"d248891604d29b0d2602e626873fba3fd54b153c6d0bc38d630237b198e9f860"} Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.982088 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" event={"ID":"ab4bae62-c176-4ff2-89af-2b9a362c4f65","Type":"ContainerStarted","Data":"a6517b03e5f1c2a909cb05f0125710405ad0b7ba2049c5df9610d72e584d7328"} Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.982505 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:45:36 crc kubenswrapper[4770]: I1209 11:45:36.999443 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-dpkk8" podStartSLOduration=1.836155856 podStartE2EDuration="4.999418447s" podCreationTimestamp="2025-12-09 11:45:32 +0000 UTC" firstStartedPulling="2025-12-09 11:45:33.054191772 +0000 UTC m=+858.294950291" lastFinishedPulling="2025-12-09 11:45:36.217454363 +0000 UTC m=+861.458212882" observedRunningTime="2025-12-09 11:45:36.991398314 +0000 UTC m=+862.232156843" watchObservedRunningTime="2025-12-09 11:45:36.999418447 +0000 UTC m=+862.240176956" Dec 09 11:45:37 crc kubenswrapper[4770]: I1209 11:45:37.018778 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" podStartSLOduration=2.014845231 podStartE2EDuration="5.018749078s" podCreationTimestamp="2025-12-09 11:45:32 +0000 UTC" firstStartedPulling="2025-12-09 11:45:33.307531301 +0000 UTC m=+858.548289820" lastFinishedPulling="2025-12-09 11:45:36.311435128 +0000 UTC m=+861.552193667" observedRunningTime="2025-12-09 11:45:37.01212072 +0000 UTC m=+862.252879249" watchObservedRunningTime="2025-12-09 11:45:37.018749078 +0000 UTC m=+862.259507597" Dec 09 11:45:40 crc kubenswrapper[4770]: I1209 11:45:40.002824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" event={"ID":"86ad5bcc-54e9-4125-a7d7-db0c74de1b01","Type":"ContainerStarted","Data":"9a0c234b92e3b1bb43c78819638ccc432e6183352233b167cbd312c206214317"} Dec 09 11:45:40 crc kubenswrapper[4770]: I1209 11:45:40.004778 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" event={"ID":"e684363b-6fd1-4432-8cee-cdb9858e4367","Type":"ContainerStarted","Data":"877fa606184684483689ca510e2ef7804baf91a13185c1dde261ab68d85a11e8"} Dec 09 11:45:40 crc kubenswrapper[4770]: I1209 11:45:40.025812 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-8s7rs" podStartSLOduration=2.1786249890000002 podStartE2EDuration="8.025784624s" podCreationTimestamp="2025-12-09 11:45:32 +0000 UTC" firstStartedPulling="2025-12-09 11:45:33.252951426 +0000 UTC m=+858.493709945" lastFinishedPulling="2025-12-09 11:45:39.100111061 +0000 UTC m=+864.340869580" observedRunningTime="2025-12-09 11:45:40.020232443 +0000 UTC m=+865.260990982" watchObservedRunningTime="2025-12-09 11:45:40.025784624 +0000 UTC m=+865.266543153" Dec 09 11:45:40 crc kubenswrapper[4770]: I1209 11:45:40.041751 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-84prv" podStartSLOduration=3.970767042 podStartE2EDuration="8.041727099s" podCreationTimestamp="2025-12-09 11:45:32 +0000 UTC" firstStartedPulling="2025-12-09 11:45:35.020834233 +0000 UTC m=+860.261592752" lastFinishedPulling="2025-12-09 11:45:39.09179429 +0000 UTC m=+864.332552809" observedRunningTime="2025-12-09 11:45:40.037786819 +0000 UTC m=+865.278545338" watchObservedRunningTime="2025-12-09 11:45:40.041727099 +0000 UTC m=+865.282485638" Dec 09 11:45:42 crc kubenswrapper[4770]: I1209 11:45:42.993442 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-dpkk8" Dec 09 11:45:43 crc kubenswrapper[4770]: I1209 11:45:43.363886 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:43 crc kubenswrapper[4770]: I1209 11:45:43.363967 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:43 crc kubenswrapper[4770]: I1209 11:45:43.364059 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:43 crc kubenswrapper[4770]: I1209 11:45:43.368849 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-68dc874465-hzpvm" Dec 09 11:45:43 crc kubenswrapper[4770]: I1209 11:45:43.451863 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:45:52 crc kubenswrapper[4770]: I1209 11:45:52.931050 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6p6bz" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.100746 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z"] Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.103000 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.105555 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.116411 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z"] Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.292372 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.292467 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.292503 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4wbz\" (UniqueName: \"kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.393986 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.394069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.394102 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4wbz\" (UniqueName: \"kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.394542 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.394652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.419761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4wbz\" (UniqueName: \"kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.423605 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.498593 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mhgdc" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" containerID="cri-o://c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383" gracePeriod=15 Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.661410 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z"] Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.855666 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mhgdc_9357bac1-cc23-4f63-8d12-458305a47f77/console/0.log" Dec 09 11:46:08 crc kubenswrapper[4770]: I1209 11:46:08.856293 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004323 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004415 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004467 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004580 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004615 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dlc6\" (UniqueName: \"kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.004636 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config\") pod \"9357bac1-cc23-4f63-8d12-458305a47f77\" (UID: \"9357bac1-cc23-4f63-8d12-458305a47f77\") " Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.005519 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config" (OuterVolumeSpecName: "console-config") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.005537 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.005567 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.005517 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca" (OuterVolumeSpecName: "service-ca") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.012411 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.012496 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.012620 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6" (OuterVolumeSpecName: "kube-api-access-2dlc6") pod "9357bac1-cc23-4f63-8d12-458305a47f77" (UID: "9357bac1-cc23-4f63-8d12-458305a47f77"). InnerVolumeSpecName "kube-api-access-2dlc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106758 4770 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106830 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106845 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106857 4770 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106869 4770 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9357bac1-cc23-4f63-8d12-458305a47f77-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106882 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dlc6\" (UniqueName: \"kubernetes.io/projected/9357bac1-cc23-4f63-8d12-458305a47f77-kube-api-access-2dlc6\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.106924 4770 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9357bac1-cc23-4f63-8d12-458305a47f77-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197399 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mhgdc_9357bac1-cc23-4f63-8d12-458305a47f77/console/0.log" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197472 4770 generic.go:334] "Generic (PLEG): container finished" podID="9357bac1-cc23-4f63-8d12-458305a47f77" containerID="c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383" exitCode=2 Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197582 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgdc" event={"ID":"9357bac1-cc23-4f63-8d12-458305a47f77","Type":"ContainerDied","Data":"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383"} Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197609 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgdc" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197713 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgdc" event={"ID":"9357bac1-cc23-4f63-8d12-458305a47f77","Type":"ContainerDied","Data":"de2c561374c6f596f237b4f07579f0d4b28314d6b3fecf04ea3ea817dcc50804"} Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.197747 4770 scope.go:117] "RemoveContainer" containerID="c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.200250 4770 generic.go:334] "Generic (PLEG): container finished" podID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerID="50de3da15f0b46e7c3eb0034f4a93d69f890f5120fc7767d7b9eea9178db45cb" exitCode=0 Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.200302 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" event={"ID":"2b7353bb-7d5f-47e4-99c5-b319206fcb90","Type":"ContainerDied","Data":"50de3da15f0b46e7c3eb0034f4a93d69f890f5120fc7767d7b9eea9178db45cb"} Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.200367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" event={"ID":"2b7353bb-7d5f-47e4-99c5-b319206fcb90","Type":"ContainerStarted","Data":"e555a601423d9f0d2ce94a9e366d70dd5f5eaa0227fc07f1f25cf01a5cf29798"} Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.220751 4770 scope.go:117] "RemoveContainer" containerID="c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383" Dec 09 11:46:09 crc kubenswrapper[4770]: E1209 11:46:09.224476 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383\": container with ID starting with c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383 not found: ID does not exist" containerID="c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.224577 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383"} err="failed to get container status \"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383\": rpc error: code = NotFound desc = could not find container \"c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383\": container with ID starting with c175da9449f22465c6281cb41868cc463b5ac8b16cc4c84adffa20a0550b4383 not found: ID does not exist" Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.247539 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.250836 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mhgdc"] Dec 09 11:46:09 crc kubenswrapper[4770]: I1209 11:46:09.349188 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" path="/var/lib/kubelet/pods/9357bac1-cc23-4f63-8d12-458305a47f77/volumes" Dec 09 11:46:11 crc kubenswrapper[4770]: I1209 11:46:11.216501 4770 generic.go:334] "Generic (PLEG): container finished" podID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerID="6fa03f7cf217c62abc7e268ab5f0ed92c3450a294e06c72e3dfef724fd4509ad" exitCode=0 Dec 09 11:46:11 crc kubenswrapper[4770]: I1209 11:46:11.216600 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" event={"ID":"2b7353bb-7d5f-47e4-99c5-b319206fcb90","Type":"ContainerDied","Data":"6fa03f7cf217c62abc7e268ab5f0ed92c3450a294e06c72e3dfef724fd4509ad"} Dec 09 11:46:12 crc kubenswrapper[4770]: I1209 11:46:12.227950 4770 generic.go:334] "Generic (PLEG): container finished" podID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerID="e0741dd1d9b6d89760b79dba7d9c523501e3c2f90f54f876e93793ef2030a426" exitCode=0 Dec 09 11:46:12 crc kubenswrapper[4770]: I1209 11:46:12.227996 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" event={"ID":"2b7353bb-7d5f-47e4-99c5-b319206fcb90","Type":"ContainerDied","Data":"e0741dd1d9b6d89760b79dba7d9c523501e3c2f90f54f876e93793ef2030a426"} Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.475848 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.574535 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4wbz\" (UniqueName: \"kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz\") pod \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.582229 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz" (OuterVolumeSpecName: "kube-api-access-c4wbz") pod "2b7353bb-7d5f-47e4-99c5-b319206fcb90" (UID: "2b7353bb-7d5f-47e4-99c5-b319206fcb90"). InnerVolumeSpecName "kube-api-access-c4wbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.675864 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util\") pod \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.676015 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle\") pod \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\" (UID: \"2b7353bb-7d5f-47e4-99c5-b319206fcb90\") " Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.676261 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4wbz\" (UniqueName: \"kubernetes.io/projected/2b7353bb-7d5f-47e4-99c5-b319206fcb90-kube-api-access-c4wbz\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.679306 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle" (OuterVolumeSpecName: "bundle") pod "2b7353bb-7d5f-47e4-99c5-b319206fcb90" (UID: "2b7353bb-7d5f-47e4-99c5-b319206fcb90"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.692405 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util" (OuterVolumeSpecName: "util") pod "2b7353bb-7d5f-47e4-99c5-b319206fcb90" (UID: "2b7353bb-7d5f-47e4-99c5-b319206fcb90"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.777394 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-util\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:13 crc kubenswrapper[4770]: I1209 11:46:13.777458 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2b7353bb-7d5f-47e4-99c5-b319206fcb90-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:14 crc kubenswrapper[4770]: I1209 11:46:14.244399 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" event={"ID":"2b7353bb-7d5f-47e4-99c5-b319206fcb90","Type":"ContainerDied","Data":"e555a601423d9f0d2ce94a9e366d70dd5f5eaa0227fc07f1f25cf01a5cf29798"} Dec 09 11:46:14 crc kubenswrapper[4770]: I1209 11:46:14.244442 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e555a601423d9f0d2ce94a9e366d70dd5f5eaa0227fc07f1f25cf01a5cf29798" Dec 09 11:46:14 crc kubenswrapper[4770]: I1209 11:46:14.244460 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.744152 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc"] Dec 09 11:46:23 crc kubenswrapper[4770]: E1209 11:46:23.745624 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.745708 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" Dec 09 11:46:23 crc kubenswrapper[4770]: E1209 11:46:23.745728 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="extract" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.745735 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="extract" Dec 09 11:46:23 crc kubenswrapper[4770]: E1209 11:46:23.745755 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="pull" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.745764 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="pull" Dec 09 11:46:23 crc kubenswrapper[4770]: E1209 11:46:23.745782 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="util" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.745789 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="util" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.746680 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7353bb-7d5f-47e4-99c5-b319206fcb90" containerName="extract" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.746717 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9357bac1-cc23-4f63-8d12-458305a47f77" containerName="console" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.761126 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.768858 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.769138 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-cdtpc" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.769416 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.769852 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.795108 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.797709 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc"] Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.942319 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m2fw\" (UniqueName: \"kubernetes.io/projected/503c0654-d161-4668-a6d5-aaca43468503-kube-api-access-8m2fw\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.942526 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-webhook-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:23 crc kubenswrapper[4770]: I1209 11:46:23.942582 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-apiservice-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.044162 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-webhook-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.044252 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-apiservice-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.044334 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m2fw\" (UniqueName: \"kubernetes.io/projected/503c0654-d161-4668-a6d5-aaca43468503-kube-api-access-8m2fw\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.049738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-apiservice-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.049738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/503c0654-d161-4668-a6d5-aaca43468503-webhook-cert\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.067377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m2fw\" (UniqueName: \"kubernetes.io/projected/503c0654-d161-4668-a6d5-aaca43468503-kube-api-access-8m2fw\") pod \"metallb-operator-controller-manager-7b5dcdff77-vbqkc\" (UID: \"503c0654-d161-4668-a6d5-aaca43468503\") " pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.092202 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb"] Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.093019 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.099547 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.099551 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.099925 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-dzkkp" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.112217 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb"] Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.115641 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.147623 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsp2g\" (UniqueName: \"kubernetes.io/projected/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-kube-api-access-zsp2g\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.147738 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-webhook-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.147766 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-apiservice-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.249410 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-webhook-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.249517 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-apiservice-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.249565 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsp2g\" (UniqueName: \"kubernetes.io/projected/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-kube-api-access-zsp2g\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.255633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-apiservice-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.259721 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-webhook-cert\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.285971 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsp2g\" (UniqueName: \"kubernetes.io/projected/13dad646-e3a3-40ea-9a0f-6892bfaae9cd-kube-api-access-zsp2g\") pod \"metallb-operator-webhook-server-7b9c5877dd-p87cb\" (UID: \"13dad646-e3a3-40ea-9a0f-6892bfaae9cd\") " pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:24 crc kubenswrapper[4770]: I1209 11:46:24.415843 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:25 crc kubenswrapper[4770]: I1209 11:46:25.070467 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb"] Dec 09 11:46:25 crc kubenswrapper[4770]: W1209 11:46:25.208114 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod503c0654_d161_4668_a6d5_aaca43468503.slice/crio-1391ff0a8c009f859a4ba26206fcc50290a0ad3cc27bf33cd03f6c1fa4cd1af8 WatchSource:0}: Error finding container 1391ff0a8c009f859a4ba26206fcc50290a0ad3cc27bf33cd03f6c1fa4cd1af8: Status 404 returned error can't find the container with id 1391ff0a8c009f859a4ba26206fcc50290a0ad3cc27bf33cd03f6c1fa4cd1af8 Dec 09 11:46:25 crc kubenswrapper[4770]: I1209 11:46:25.214870 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc"] Dec 09 11:46:25 crc kubenswrapper[4770]: I1209 11:46:25.356972 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" event={"ID":"503c0654-d161-4668-a6d5-aaca43468503","Type":"ContainerStarted","Data":"1391ff0a8c009f859a4ba26206fcc50290a0ad3cc27bf33cd03f6c1fa4cd1af8"} Dec 09 11:46:25 crc kubenswrapper[4770]: I1209 11:46:25.358729 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" event={"ID":"13dad646-e3a3-40ea-9a0f-6892bfaae9cd","Type":"ContainerStarted","Data":"2dea196dae5aa5a988405faaa4428b0a90d4bb12ea9f08dcbc76bb4695c583d4"} Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.747230 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.749244 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.760280 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.895511 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.896143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.896528 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ctb\" (UniqueName: \"kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.998574 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ctb\" (UniqueName: \"kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.998648 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.998670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.999504 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:32 crc kubenswrapper[4770]: I1209 11:46:32.999551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:33 crc kubenswrapper[4770]: I1209 11:46:33.023255 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ctb\" (UniqueName: \"kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb\") pod \"certified-operators-djsjh\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:33 crc kubenswrapper[4770]: I1209 11:46:33.096264 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:33 crc kubenswrapper[4770]: I1209 11:46:33.966404 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:33 crc kubenswrapper[4770]: W1209 11:46:33.979375 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod656fca69_3db4_44ee_9960_e56cd44a2e77.slice/crio-bdb14713e7c9810b50da1a0c491845347b3b99bf2d8a694cc2077bdb64646295 WatchSource:0}: Error finding container bdb14713e7c9810b50da1a0c491845347b3b99bf2d8a694cc2077bdb64646295: Status 404 returned error can't find the container with id bdb14713e7c9810b50da1a0c491845347b3b99bf2d8a694cc2077bdb64646295 Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.513221 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" event={"ID":"503c0654-d161-4668-a6d5-aaca43468503","Type":"ContainerStarted","Data":"4e4ac1d130688d95a8a16b704fada5d71b501af0b62a50165f67d92ff4eccc3c"} Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.513725 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.515724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" event={"ID":"13dad646-e3a3-40ea-9a0f-6892bfaae9cd","Type":"ContainerStarted","Data":"a6d184309d2c4d2dd3636f7c9e1e64aaaade9c00bc4affa68ac820579d64bcc4"} Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.515887 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.518194 4770 generic.go:334] "Generic (PLEG): container finished" podID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerID="dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb" exitCode=0 Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.518254 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerDied","Data":"dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb"} Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.518289 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerStarted","Data":"bdb14713e7c9810b50da1a0c491845347b3b99bf2d8a694cc2077bdb64646295"} Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.544084 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" podStartSLOduration=3.548853487 podStartE2EDuration="11.544066998s" podCreationTimestamp="2025-12-09 11:46:23 +0000 UTC" firstStartedPulling="2025-12-09 11:46:25.211640331 +0000 UTC m=+910.452398860" lastFinishedPulling="2025-12-09 11:46:33.206853852 +0000 UTC m=+918.447612371" observedRunningTime="2025-12-09 11:46:34.540788655 +0000 UTC m=+919.781547214" watchObservedRunningTime="2025-12-09 11:46:34.544066998 +0000 UTC m=+919.784825517" Dec 09 11:46:34 crc kubenswrapper[4770]: I1209 11:46:34.566889 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" podStartSLOduration=2.4406698430000002 podStartE2EDuration="10.566852257s" podCreationTimestamp="2025-12-09 11:46:24 +0000 UTC" firstStartedPulling="2025-12-09 11:46:25.082543265 +0000 UTC m=+910.323301784" lastFinishedPulling="2025-12-09 11:46:33.208725679 +0000 UTC m=+918.449484198" observedRunningTime="2025-12-09 11:46:34.560450814 +0000 UTC m=+919.801209323" watchObservedRunningTime="2025-12-09 11:46:34.566852257 +0000 UTC m=+919.807610776" Dec 09 11:46:36 crc kubenswrapper[4770]: I1209 11:46:36.536764 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerStarted","Data":"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9"} Dec 09 11:46:37 crc kubenswrapper[4770]: I1209 11:46:37.547862 4770 generic.go:334] "Generic (PLEG): container finished" podID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerID="abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9" exitCode=0 Dec 09 11:46:37 crc kubenswrapper[4770]: I1209 11:46:37.548170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerDied","Data":"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9"} Dec 09 11:46:38 crc kubenswrapper[4770]: I1209 11:46:38.558641 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerStarted","Data":"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d"} Dec 09 11:46:38 crc kubenswrapper[4770]: I1209 11:46:38.582058 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-djsjh" podStartSLOduration=3.125254758 podStartE2EDuration="6.582034348s" podCreationTimestamp="2025-12-09 11:46:32 +0000 UTC" firstStartedPulling="2025-12-09 11:46:34.520076149 +0000 UTC m=+919.760834668" lastFinishedPulling="2025-12-09 11:46:37.976855739 +0000 UTC m=+923.217614258" observedRunningTime="2025-12-09 11:46:38.579869793 +0000 UTC m=+923.820628332" watchObservedRunningTime="2025-12-09 11:46:38.582034348 +0000 UTC m=+923.822792877" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.115254 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.116876 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.125002 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.282418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9btnj\" (UniqueName: \"kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.282488 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.282551 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.384237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.384361 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9btnj\" (UniqueName: \"kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.384427 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.385023 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.385240 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.408446 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9btnj\" (UniqueName: \"kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj\") pod \"redhat-marketplace-nvktd\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:42 crc kubenswrapper[4770]: I1209 11:46:42.448406 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.099270 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.100250 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.233858 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.424353 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.591859 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerStarted","Data":"bcd54d1a0e79a322ea06cc261c8db9d6a740974f4eb9dc0bab9f3e9a626c7cec"} Dec 09 11:46:43 crc kubenswrapper[4770]: I1209 11:46:43.640479 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:44 crc kubenswrapper[4770]: I1209 11:46:44.424466 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7b9c5877dd-p87cb" Dec 09 11:46:44 crc kubenswrapper[4770]: I1209 11:46:44.600132 4770 generic.go:334] "Generic (PLEG): container finished" podID="7037efca-c072-44ba-9294-2177c9820445" containerID="716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d" exitCode=0 Dec 09 11:46:44 crc kubenswrapper[4770]: I1209 11:46:44.601371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerDied","Data":"716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d"} Dec 09 11:46:45 crc kubenswrapper[4770]: I1209 11:46:45.806236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerStarted","Data":"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce"} Dec 09 11:46:46 crc kubenswrapper[4770]: I1209 11:46:46.815081 4770 generic.go:334] "Generic (PLEG): container finished" podID="7037efca-c072-44ba-9294-2177c9820445" containerID="639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce" exitCode=0 Dec 09 11:46:46 crc kubenswrapper[4770]: I1209 11:46:46.815155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerDied","Data":"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce"} Dec 09 11:46:46 crc kubenswrapper[4770]: I1209 11:46:46.902021 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:46 crc kubenswrapper[4770]: I1209 11:46:46.902268 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-djsjh" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="registry-server" containerID="cri-o://9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d" gracePeriod=2 Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.799703 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.831969 4770 generic.go:334] "Generic (PLEG): container finished" podID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerID="9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d" exitCode=0 Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.832105 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerDied","Data":"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d"} Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.832150 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsjh" event={"ID":"656fca69-3db4-44ee-9960-e56cd44a2e77","Type":"ContainerDied","Data":"bdb14713e7c9810b50da1a0c491845347b3b99bf2d8a694cc2077bdb64646295"} Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.832195 4770 scope.go:117] "RemoveContainer" containerID="9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.832435 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsjh" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.846087 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerStarted","Data":"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0"} Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.860577 4770 scope.go:117] "RemoveContainer" containerID="abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.868003 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nvktd" podStartSLOduration=2.999603971 podStartE2EDuration="5.867983867s" podCreationTimestamp="2025-12-09 11:46:42 +0000 UTC" firstStartedPulling="2025-12-09 11:46:44.602180494 +0000 UTC m=+929.842939013" lastFinishedPulling="2025-12-09 11:46:47.47056039 +0000 UTC m=+932.711318909" observedRunningTime="2025-12-09 11:46:47.865179225 +0000 UTC m=+933.105937744" watchObservedRunningTime="2025-12-09 11:46:47.867983867 +0000 UTC m=+933.108742386" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.889626 4770 scope.go:117] "RemoveContainer" containerID="dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.906597 4770 scope.go:117] "RemoveContainer" containerID="9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d" Dec 09 11:46:47 crc kubenswrapper[4770]: E1209 11:46:47.907465 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d\": container with ID starting with 9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d not found: ID does not exist" containerID="9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.907538 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d"} err="failed to get container status \"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d\": rpc error: code = NotFound desc = could not find container \"9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d\": container with ID starting with 9bd24b4dfbc502cdf62dc86615988cb92620683579598b7ce8414682aafe6a9d not found: ID does not exist" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.907646 4770 scope.go:117] "RemoveContainer" containerID="abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9" Dec 09 11:46:47 crc kubenswrapper[4770]: E1209 11:46:47.908087 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9\": container with ID starting with abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9 not found: ID does not exist" containerID="abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.908122 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9"} err="failed to get container status \"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9\": rpc error: code = NotFound desc = could not find container \"abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9\": container with ID starting with abb0dda190a04a6b4a0d0d0bb12df05c2eaa798d2367a884cf9cca36270dcbd9 not found: ID does not exist" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.908145 4770 scope.go:117] "RemoveContainer" containerID="dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb" Dec 09 11:46:47 crc kubenswrapper[4770]: E1209 11:46:47.908802 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb\": container with ID starting with dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb not found: ID does not exist" containerID="dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.908835 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb"} err="failed to get container status \"dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb\": rpc error: code = NotFound desc = could not find container \"dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb\": container with ID starting with dae5f9ca1282e5de68ef3ad4ff9b7890d3586976bd9b377b61d24b2bc04dbebb not found: ID does not exist" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.931861 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content\") pod \"656fca69-3db4-44ee-9960-e56cd44a2e77\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.932032 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ctb\" (UniqueName: \"kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb\") pod \"656fca69-3db4-44ee-9960-e56cd44a2e77\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.932079 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities\") pod \"656fca69-3db4-44ee-9960-e56cd44a2e77\" (UID: \"656fca69-3db4-44ee-9960-e56cd44a2e77\") " Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.935680 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities" (OuterVolumeSpecName: "utilities") pod "656fca69-3db4-44ee-9960-e56cd44a2e77" (UID: "656fca69-3db4-44ee-9960-e56cd44a2e77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.942065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb" (OuterVolumeSpecName: "kube-api-access-29ctb") pod "656fca69-3db4-44ee-9960-e56cd44a2e77" (UID: "656fca69-3db4-44ee-9960-e56cd44a2e77"). InnerVolumeSpecName "kube-api-access-29ctb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:46:47 crc kubenswrapper[4770]: I1209 11:46:47.992540 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "656fca69-3db4-44ee-9960-e56cd44a2e77" (UID: "656fca69-3db4-44ee-9960-e56cd44a2e77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:48 crc kubenswrapper[4770]: I1209 11:46:48.035396 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:48 crc kubenswrapper[4770]: I1209 11:46:48.035936 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ctb\" (UniqueName: \"kubernetes.io/projected/656fca69-3db4-44ee-9960-e56cd44a2e77-kube-api-access-29ctb\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:48 crc kubenswrapper[4770]: I1209 11:46:48.035975 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656fca69-3db4-44ee-9960-e56cd44a2e77-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:48 crc kubenswrapper[4770]: I1209 11:46:48.162971 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:48 crc kubenswrapper[4770]: I1209 11:46:48.167472 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-djsjh"] Dec 09 11:46:49 crc kubenswrapper[4770]: I1209 11:46:49.350158 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" path="/var/lib/kubelet/pods/656fca69-3db4-44ee-9960-e56cd44a2e77/volumes" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.510145 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:46:51 crc kubenswrapper[4770]: E1209 11:46:51.510527 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="extract-utilities" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.510548 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="extract-utilities" Dec 09 11:46:51 crc kubenswrapper[4770]: E1209 11:46:51.510577 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="extract-content" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.510585 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="extract-content" Dec 09 11:46:51 crc kubenswrapper[4770]: E1209 11:46:51.510599 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="registry-server" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.510608 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="registry-server" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.510760 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="656fca69-3db4-44ee-9960-e56cd44a2e77" containerName="registry-server" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.511948 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.524065 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.585818 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkxgc\" (UniqueName: \"kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.586014 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.586048 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.687567 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.687627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.687696 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkxgc\" (UniqueName: \"kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.688296 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.688291 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.720989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkxgc\" (UniqueName: \"kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc\") pod \"community-operators-4gxnj\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:51 crc kubenswrapper[4770]: I1209 11:46:51.829417 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.393720 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.451275 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.451349 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.518807 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.887497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerStarted","Data":"4d2b650a9396c0cf8919dfa1c142c08bd89f09541c37c5e3bd7f797251aaebc2"} Dec 09 11:46:52 crc kubenswrapper[4770]: I1209 11:46:52.935010 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:53 crc kubenswrapper[4770]: I1209 11:46:53.897569 4770 generic.go:334] "Generic (PLEG): container finished" podID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerID="8055c142fcbd2e2b772fe861c51b1f691b6a97f787a75a50637337a3a1bd628f" exitCode=0 Dec 09 11:46:53 crc kubenswrapper[4770]: I1209 11:46:53.897642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerDied","Data":"8055c142fcbd2e2b772fe861c51b1f691b6a97f787a75a50637337a3a1bd628f"} Dec 09 11:46:54 crc kubenswrapper[4770]: I1209 11:46:54.905023 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerStarted","Data":"5fc05eedf6ab6cb5b4a7b3841b04d456875def9e43d90c0ca4e76860f2fa00f2"} Dec 09 11:46:55 crc kubenswrapper[4770]: I1209 11:46:55.901062 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:55 crc kubenswrapper[4770]: I1209 11:46:55.901311 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nvktd" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="registry-server" containerID="cri-o://f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0" gracePeriod=2 Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.811021 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.913055 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content\") pod \"7037efca-c072-44ba-9294-2177c9820445\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.913116 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities\") pod \"7037efca-c072-44ba-9294-2177c9820445\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.913169 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9btnj\" (UniqueName: \"kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj\") pod \"7037efca-c072-44ba-9294-2177c9820445\" (UID: \"7037efca-c072-44ba-9294-2177c9820445\") " Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.914571 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities" (OuterVolumeSpecName: "utilities") pod "7037efca-c072-44ba-9294-2177c9820445" (UID: "7037efca-c072-44ba-9294-2177c9820445"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.918429 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj" (OuterVolumeSpecName: "kube-api-access-9btnj") pod "7037efca-c072-44ba-9294-2177c9820445" (UID: "7037efca-c072-44ba-9294-2177c9820445"). InnerVolumeSpecName "kube-api-access-9btnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.923991 4770 generic.go:334] "Generic (PLEG): container finished" podID="7037efca-c072-44ba-9294-2177c9820445" containerID="f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0" exitCode=0 Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.924107 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvktd" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.924649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerDied","Data":"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0"} Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.925020 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvktd" event={"ID":"7037efca-c072-44ba-9294-2177c9820445","Type":"ContainerDied","Data":"bcd54d1a0e79a322ea06cc261c8db9d6a740974f4eb9dc0bab9f3e9a626c7cec"} Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.925154 4770 scope.go:117] "RemoveContainer" containerID="f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.926990 4770 generic.go:334] "Generic (PLEG): container finished" podID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerID="5fc05eedf6ab6cb5b4a7b3841b04d456875def9e43d90c0ca4e76860f2fa00f2" exitCode=0 Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.927032 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerDied","Data":"5fc05eedf6ab6cb5b4a7b3841b04d456875def9e43d90c0ca4e76860f2fa00f2"} Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.935225 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7037efca-c072-44ba-9294-2177c9820445" (UID: "7037efca-c072-44ba-9294-2177c9820445"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.958972 4770 scope.go:117] "RemoveContainer" containerID="639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce" Dec 09 11:46:56 crc kubenswrapper[4770]: I1209 11:46:56.984266 4770 scope.go:117] "RemoveContainer" containerID="716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.010104 4770 scope.go:117] "RemoveContainer" containerID="f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0" Dec 09 11:46:57 crc kubenswrapper[4770]: E1209 11:46:57.010659 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0\": container with ID starting with f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0 not found: ID does not exist" containerID="f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.010692 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0"} err="failed to get container status \"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0\": rpc error: code = NotFound desc = could not find container \"f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0\": container with ID starting with f25ad5bfe9b7754fdaa9faaead885efb0fbb6feafa88d4f6415567ffeaa616d0 not found: ID does not exist" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.010717 4770 scope.go:117] "RemoveContainer" containerID="639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce" Dec 09 11:46:57 crc kubenswrapper[4770]: E1209 11:46:57.011315 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce\": container with ID starting with 639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce not found: ID does not exist" containerID="639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.011343 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce"} err="failed to get container status \"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce\": rpc error: code = NotFound desc = could not find container \"639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce\": container with ID starting with 639d36a871929acb2616d9df4f8083db631c91b405cc97acdc4342c4811c7bce not found: ID does not exist" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.011359 4770 scope.go:117] "RemoveContainer" containerID="716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d" Dec 09 11:46:57 crc kubenswrapper[4770]: E1209 11:46:57.011702 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d\": container with ID starting with 716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d not found: ID does not exist" containerID="716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.011724 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d"} err="failed to get container status \"716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d\": rpc error: code = NotFound desc = could not find container \"716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d\": container with ID starting with 716e3f8dfbbf79bf554632198243ca989435e6becc64d97c1b9d39ad48d8c76d not found: ID does not exist" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.014551 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.014589 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7037efca-c072-44ba-9294-2177c9820445-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.014639 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9btnj\" (UniqueName: \"kubernetes.io/projected/7037efca-c072-44ba-9294-2177c9820445-kube-api-access-9btnj\") on node \"crc\" DevicePath \"\"" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.254587 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.258614 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvktd"] Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.349916 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7037efca-c072-44ba-9294-2177c9820445" path="/var/lib/kubelet/pods/7037efca-c072-44ba-9294-2177c9820445/volumes" Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.937556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerStarted","Data":"0bcbbb9c38b7fe30203b728a16543b721cccedae5adc679e779220bf6b159e85"} Dec 09 11:46:57 crc kubenswrapper[4770]: I1209 11:46:57.960152 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4gxnj" podStartSLOduration=3.465987987 podStartE2EDuration="6.960135135s" podCreationTimestamp="2025-12-09 11:46:51 +0000 UTC" firstStartedPulling="2025-12-09 11:46:53.899802887 +0000 UTC m=+939.140561406" lastFinishedPulling="2025-12-09 11:46:57.393950035 +0000 UTC m=+942.634708554" observedRunningTime="2025-12-09 11:46:57.956198935 +0000 UTC m=+943.196957454" watchObservedRunningTime="2025-12-09 11:46:57.960135135 +0000 UTC m=+943.200893654" Dec 09 11:47:01 crc kubenswrapper[4770]: I1209 11:47:01.829621 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:01 crc kubenswrapper[4770]: I1209 11:47:01.830250 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:01 crc kubenswrapper[4770]: I1209 11:47:01.873176 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:02 crc kubenswrapper[4770]: I1209 11:47:02.473888 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:47:02 crc kubenswrapper[4770]: I1209 11:47:02.474006 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.120171 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7b5dcdff77-vbqkc" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.882738 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-8b799"] Dec 09 11:47:04 crc kubenswrapper[4770]: E1209 11:47:04.883022 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="extract-content" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.883039 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="extract-content" Dec 09 11:47:04 crc kubenswrapper[4770]: E1209 11:47:04.883059 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="extract-utilities" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.883066 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="extract-utilities" Dec 09 11:47:04 crc kubenswrapper[4770]: E1209 11:47:04.883078 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="registry-server" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.883084 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="registry-server" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.883210 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7037efca-c072-44ba-9294-2177c9820445" containerName="registry-server" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.885881 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.889475 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.889789 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.893003 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-r7z72" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.895566 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84"] Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.896454 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.898941 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 09 11:47:04 crc kubenswrapper[4770]: I1209 11:47:04.913003 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84"] Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.049618 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7kzks"] Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.050608 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.054621 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bbx49" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.055115 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.057646 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.060459 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067613 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067676 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067702 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-startup\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067728 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b906317-0543-467d-9e51-a3af3f40d6b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067751 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-reloader\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067766 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067781 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics-certs\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067801 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-sockets\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7jvl\" (UniqueName: \"kubernetes.io/projected/4b906317-0543-467d-9e51-a3af3f40d6b6-kube-api-access-k7jvl\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067844 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-conf\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxbfh\" (UniqueName: \"kubernetes.io/projected/4002944e-6d2f-4f2b-8d01-e664da3116e2-kube-api-access-sxbfh\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067935 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7pb5\" (UniqueName: \"kubernetes.io/projected/e013d4c8-5587-407a-9a8f-b26ab7064ec8-kube-api-access-p7pb5\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.067961 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metallb-excludel2\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.109189 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-n9pcb"] Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.110244 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.113366 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.156391 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-n9pcb"] Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169652 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxbfh\" (UniqueName: \"kubernetes.io/projected/4002944e-6d2f-4f2b-8d01-e664da3116e2-kube-api-access-sxbfh\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169772 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-cert\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7pb5\" (UniqueName: \"kubernetes.io/projected/e013d4c8-5587-407a-9a8f-b26ab7064ec8-kube-api-access-p7pb5\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metallb-excludel2\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169878 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gh5w\" (UniqueName: \"kubernetes.io/projected/844a755c-0f78-40a6-93f3-a388dce59835-kube-api-access-8gh5w\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169964 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.169987 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-startup\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170010 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b906317-0543-467d-9e51-a3af3f40d6b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170028 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-reloader\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170047 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170068 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics-certs\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170088 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-sockets\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170109 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7jvl\" (UniqueName: \"kubernetes.io/projected/4b906317-0543-467d-9e51-a3af3f40d6b6-kube-api-access-k7jvl\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170133 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-conf\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.170690 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-conf\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.172068 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metallb-excludel2\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.172322 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-reloader\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.172514 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.173325 4770 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.173392 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist podName:e013d4c8-5587-407a-9a8f-b26ab7064ec8 nodeName:}" failed. No retries permitted until 2025-12-09 11:47:05.673372629 +0000 UTC m=+950.914131148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist") pod "speaker-7kzks" (UID: "e013d4c8-5587-407a-9a8f-b26ab7064ec8") : secret "metallb-memberlist" not found Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.173681 4770 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.173721 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs podName:e013d4c8-5587-407a-9a8f-b26ab7064ec8 nodeName:}" failed. No retries permitted until 2025-12-09 11:47:05.673711978 +0000 UTC m=+950.914470497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs") pod "speaker-7kzks" (UID: "e013d4c8-5587-407a-9a8f-b26ab7064ec8") : secret "speaker-certs-secret" not found Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.174236 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-sockets\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.174726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4002944e-6d2f-4f2b-8d01-e664da3116e2-frr-startup\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.180610 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4002944e-6d2f-4f2b-8d01-e664da3116e2-metrics-certs\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.196658 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b906317-0543-467d-9e51-a3af3f40d6b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.213684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxbfh\" (UniqueName: \"kubernetes.io/projected/4002944e-6d2f-4f2b-8d01-e664da3116e2-kube-api-access-sxbfh\") pod \"frr-k8s-8b799\" (UID: \"4002944e-6d2f-4f2b-8d01-e664da3116e2\") " pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.218701 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7pb5\" (UniqueName: \"kubernetes.io/projected/e013d4c8-5587-407a-9a8f-b26ab7064ec8-kube-api-access-p7pb5\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.229457 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.234665 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7jvl\" (UniqueName: \"kubernetes.io/projected/4b906317-0543-467d-9e51-a3af3f40d6b6-kube-api-access-k7jvl\") pod \"frr-k8s-webhook-server-7fcb986d4-jlj84\" (UID: \"4b906317-0543-467d-9e51-a3af3f40d6b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.270841 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.270893 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-cert\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.270949 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gh5w\" (UniqueName: \"kubernetes.io/projected/844a755c-0f78-40a6-93f3-a388dce59835-kube-api-access-8gh5w\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.271079 4770 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.271142 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs podName:844a755c-0f78-40a6-93f3-a388dce59835 nodeName:}" failed. No retries permitted until 2025-12-09 11:47:05.77112325 +0000 UTC m=+951.011881769 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs") pod "controller-f8648f98b-n9pcb" (UID: "844a755c-0f78-40a6-93f3-a388dce59835") : secret "controller-certs-secret" not found Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.274943 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-cert\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.293225 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gh5w\" (UniqueName: \"kubernetes.io/projected/844a755c-0f78-40a6-93f3-a388dce59835-kube-api-access-8gh5w\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.525380 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.682376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.683068 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.683251 4770 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 11:47:05 crc kubenswrapper[4770]: E1209 11:47:05.683333 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist podName:e013d4c8-5587-407a-9a8f-b26ab7064ec8 nodeName:}" failed. No retries permitted until 2025-12-09 11:47:06.683307151 +0000 UTC m=+951.924065670 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist") pod "speaker-7kzks" (UID: "e013d4c8-5587-407a-9a8f-b26ab7064ec8") : secret "metallb-memberlist" not found Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.700158 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-metrics-certs\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.746616 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84"] Dec 09 11:47:05 crc kubenswrapper[4770]: W1209 11:47:05.756771 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b906317_0543_467d_9e51_a3af3f40d6b6.slice/crio-6be2a87558247cbf968f0d3b8dbaa4ff5bf817f997cf8882f5ef2c11f77d88b6 WatchSource:0}: Error finding container 6be2a87558247cbf968f0d3b8dbaa4ff5bf817f997cf8882f5ef2c11f77d88b6: Status 404 returned error can't find the container with id 6be2a87558247cbf968f0d3b8dbaa4ff5bf817f997cf8882f5ef2c11f77d88b6 Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.784177 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.789540 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/844a755c-0f78-40a6-93f3-a388dce59835-metrics-certs\") pod \"controller-f8648f98b-n9pcb\" (UID: \"844a755c-0f78-40a6-93f3-a388dce59835\") " pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.985672 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"6b317c6a3a77e08a67511c48999f66c4108981dde68447290db4a4156a7992d9"} Dec 09 11:47:05 crc kubenswrapper[4770]: I1209 11:47:05.986562 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" event={"ID":"4b906317-0543-467d-9e51-a3af3f40d6b6","Type":"ContainerStarted","Data":"6be2a87558247cbf968f0d3b8dbaa4ff5bf817f997cf8882f5ef2c11f77d88b6"} Dec 09 11:47:06 crc kubenswrapper[4770]: I1209 11:47:06.027996 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:06 crc kubenswrapper[4770]: I1209 11:47:06.254178 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-n9pcb"] Dec 09 11:47:06 crc kubenswrapper[4770]: I1209 11:47:06.704259 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:06 crc kubenswrapper[4770]: I1209 11:47:06.710924 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e013d4c8-5587-407a-9a8f-b26ab7064ec8-memberlist\") pod \"speaker-7kzks\" (UID: \"e013d4c8-5587-407a-9a8f-b26ab7064ec8\") " pod="metallb-system/speaker-7kzks" Dec 09 11:47:06 crc kubenswrapper[4770]: I1209 11:47:06.864243 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7kzks" Dec 09 11:47:06 crc kubenswrapper[4770]: W1209 11:47:06.896997 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode013d4c8_5587_407a_9a8f_b26ab7064ec8.slice/crio-d0f4449039f64990c2ed8e716f83e33edbc11565d3182464e689c2a3a40e9049 WatchSource:0}: Error finding container d0f4449039f64990c2ed8e716f83e33edbc11565d3182464e689c2a3a40e9049: Status 404 returned error can't find the container with id d0f4449039f64990c2ed8e716f83e33edbc11565d3182464e689c2a3a40e9049 Dec 09 11:47:07 crc kubenswrapper[4770]: I1209 11:47:07.001502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n9pcb" event={"ID":"844a755c-0f78-40a6-93f3-a388dce59835","Type":"ContainerStarted","Data":"0689b44957ef2495990d11d4cacfd6ce1865d73f636b3ccced70b1f6b042d358"} Dec 09 11:47:07 crc kubenswrapper[4770]: I1209 11:47:07.001554 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n9pcb" event={"ID":"844a755c-0f78-40a6-93f3-a388dce59835","Type":"ContainerStarted","Data":"f93127c93562ff1b8fa3a5fa9ce595056f8152b295e9218549a15e41e8c07a46"} Dec 09 11:47:07 crc kubenswrapper[4770]: I1209 11:47:07.001564 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n9pcb" event={"ID":"844a755c-0f78-40a6-93f3-a388dce59835","Type":"ContainerStarted","Data":"d948394f524ada1989b3d6f817f06132f63be106bdc50aa485a37ec4aa9ec4b7"} Dec 09 11:47:07 crc kubenswrapper[4770]: I1209 11:47:07.001682 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:07 crc kubenswrapper[4770]: I1209 11:47:07.003073 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7kzks" event={"ID":"e013d4c8-5587-407a-9a8f-b26ab7064ec8","Type":"ContainerStarted","Data":"d0f4449039f64990c2ed8e716f83e33edbc11565d3182464e689c2a3a40e9049"} Dec 09 11:47:08 crc kubenswrapper[4770]: I1209 11:47:08.079870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7kzks" event={"ID":"e013d4c8-5587-407a-9a8f-b26ab7064ec8","Type":"ContainerStarted","Data":"8454ae8152570e4c5f23f60fe29b1c67074f021d3e81b1c6bde02cfe56816f0b"} Dec 09 11:47:08 crc kubenswrapper[4770]: I1209 11:47:08.080445 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7kzks" event={"ID":"e013d4c8-5587-407a-9a8f-b26ab7064ec8","Type":"ContainerStarted","Data":"cc1ae6e85bcb2cdd56dda65f1d173f07ae337284846b26e66fece70d5efdd253"} Dec 09 11:47:08 crc kubenswrapper[4770]: I1209 11:47:08.080488 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7kzks" Dec 09 11:47:08 crc kubenswrapper[4770]: I1209 11:47:08.113231 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-n9pcb" podStartSLOduration=3.113197208 podStartE2EDuration="3.113197208s" podCreationTimestamp="2025-12-09 11:47:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:47:07.059301542 +0000 UTC m=+952.300060061" watchObservedRunningTime="2025-12-09 11:47:08.113197208 +0000 UTC m=+953.353955717" Dec 09 11:47:08 crc kubenswrapper[4770]: I1209 11:47:08.114699 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7kzks" podStartSLOduration=3.114685096 podStartE2EDuration="3.114685096s" podCreationTimestamp="2025-12-09 11:47:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:47:08.108992812 +0000 UTC m=+953.349751331" watchObservedRunningTime="2025-12-09 11:47:08.114685096 +0000 UTC m=+953.355443635" Dec 09 11:47:11 crc kubenswrapper[4770]: I1209 11:47:11.937883 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:11 crc kubenswrapper[4770]: I1209 11:47:11.995675 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:47:12 crc kubenswrapper[4770]: I1209 11:47:12.116847 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4gxnj" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="registry-server" containerID="cri-o://0bcbbb9c38b7fe30203b728a16543b721cccedae5adc679e779220bf6b159e85" gracePeriod=2 Dec 09 11:47:13 crc kubenswrapper[4770]: I1209 11:47:13.144360 4770 generic.go:334] "Generic (PLEG): container finished" podID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerID="0bcbbb9c38b7fe30203b728a16543b721cccedae5adc679e779220bf6b159e85" exitCode=0 Dec 09 11:47:13 crc kubenswrapper[4770]: I1209 11:47:13.145332 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerDied","Data":"0bcbbb9c38b7fe30203b728a16543b721cccedae5adc679e779220bf6b159e85"} Dec 09 11:47:16 crc kubenswrapper[4770]: I1209 11:47:16.034517 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-n9pcb" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.758127 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.865972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities\") pod \"40f6fe58-d863-4a6c-8721-d9d721a21810\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.866107 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkxgc\" (UniqueName: \"kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc\") pod \"40f6fe58-d863-4a6c-8721-d9d721a21810\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.866195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content\") pod \"40f6fe58-d863-4a6c-8721-d9d721a21810\" (UID: \"40f6fe58-d863-4a6c-8721-d9d721a21810\") " Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.866804 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities" (OuterVolumeSpecName: "utilities") pod "40f6fe58-d863-4a6c-8721-d9d721a21810" (UID: "40f6fe58-d863-4a6c-8721-d9d721a21810"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.888133 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc" (OuterVolumeSpecName: "kube-api-access-qkxgc") pod "40f6fe58-d863-4a6c-8721-d9d721a21810" (UID: "40f6fe58-d863-4a6c-8721-d9d721a21810"). InnerVolumeSpecName "kube-api-access-qkxgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.929748 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40f6fe58-d863-4a6c-8721-d9d721a21810" (UID: "40f6fe58-d863-4a6c-8721-d9d721a21810"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.967419 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkxgc\" (UniqueName: \"kubernetes.io/projected/40f6fe58-d863-4a6c-8721-d9d721a21810-kube-api-access-qkxgc\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.967461 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:17 crc kubenswrapper[4770]: I1209 11:47:17.967472 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f6fe58-d863-4a6c-8721-d9d721a21810-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.185699 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gxnj" event={"ID":"40f6fe58-d863-4a6c-8721-d9d721a21810","Type":"ContainerDied","Data":"4d2b650a9396c0cf8919dfa1c142c08bd89f09541c37c5e3bd7f797251aaebc2"} Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.185758 4770 scope.go:117] "RemoveContainer" containerID="0bcbbb9c38b7fe30203b728a16543b721cccedae5adc679e779220bf6b159e85" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.185869 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gxnj" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.190226 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" event={"ID":"4b906317-0543-467d-9e51-a3af3f40d6b6","Type":"ContainerStarted","Data":"eaadf64a94309282882a1d2bf989707cacbd8f669d278da2a093c62eae73982d"} Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.190807 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.192146 4770 generic.go:334] "Generic (PLEG): container finished" podID="4002944e-6d2f-4f2b-8d01-e664da3116e2" containerID="36bca508dd942c68fa7319862f34e42661ea1e4dee7c2d42e4a7da72b92cb2d8" exitCode=0 Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.192212 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerDied","Data":"36bca508dd942c68fa7319862f34e42661ea1e4dee7c2d42e4a7da72b92cb2d8"} Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.204969 4770 scope.go:117] "RemoveContainer" containerID="5fc05eedf6ab6cb5b4a7b3841b04d456875def9e43d90c0ca4e76860f2fa00f2" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.235685 4770 scope.go:117] "RemoveContainer" containerID="8055c142fcbd2e2b772fe861c51b1f691b6a97f787a75a50637337a3a1bd628f" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.242353 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" podStartSLOduration=2.428075002 podStartE2EDuration="14.242303665s" podCreationTimestamp="2025-12-09 11:47:04 +0000 UTC" firstStartedPulling="2025-12-09 11:47:05.76405943 +0000 UTC m=+951.004817949" lastFinishedPulling="2025-12-09 11:47:17.578288093 +0000 UTC m=+962.819046612" observedRunningTime="2025-12-09 11:47:18.235835741 +0000 UTC m=+963.476594260" watchObservedRunningTime="2025-12-09 11:47:18.242303665 +0000 UTC m=+963.483062184" Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.258729 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:47:18 crc kubenswrapper[4770]: I1209 11:47:18.262995 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4gxnj"] Dec 09 11:47:19 crc kubenswrapper[4770]: I1209 11:47:19.201357 4770 generic.go:334] "Generic (PLEG): container finished" podID="4002944e-6d2f-4f2b-8d01-e664da3116e2" containerID="8ff2aba6da34e419341b9b656c266f8d0692601f28b9fcd7317ddb55136bce16" exitCode=0 Dec 09 11:47:19 crc kubenswrapper[4770]: I1209 11:47:19.201842 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerDied","Data":"8ff2aba6da34e419341b9b656c266f8d0692601f28b9fcd7317ddb55136bce16"} Dec 09 11:47:19 crc kubenswrapper[4770]: I1209 11:47:19.350102 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" path="/var/lib/kubelet/pods/40f6fe58-d863-4a6c-8721-d9d721a21810/volumes" Dec 09 11:47:20 crc kubenswrapper[4770]: I1209 11:47:20.214835 4770 generic.go:334] "Generic (PLEG): container finished" podID="4002944e-6d2f-4f2b-8d01-e664da3116e2" containerID="ab73a3d5a9dd19a965a7891e584cce04e749114918bf9fe26f23c019e0eb2861" exitCode=0 Dec 09 11:47:20 crc kubenswrapper[4770]: I1209 11:47:20.214989 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerDied","Data":"ab73a3d5a9dd19a965a7891e584cce04e749114918bf9fe26f23c019e0eb2861"} Dec 09 11:47:21 crc kubenswrapper[4770]: I1209 11:47:21.227081 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"be1b8df70585a017952b2487bc7824e2d9afbb71235417e241ff3058d93ae222"} Dec 09 11:47:22 crc kubenswrapper[4770]: I1209 11:47:22.246835 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"7cded3b8886d34e7fe7b144a90ac222472f2ed6d92c90f7f5bac752b1067ebeb"} Dec 09 11:47:22 crc kubenswrapper[4770]: I1209 11:47:22.247391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"c0d5dcffad3f6ee21d82c08a6833d6f495b7f7cbc4aeda4d1c2b3252de727c37"} Dec 09 11:47:22 crc kubenswrapper[4770]: I1209 11:47:22.247407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"168752a63f107a461f748f4ab59ead55a3420f18fdbf415b4691af31773824e3"} Dec 09 11:47:23 crc kubenswrapper[4770]: I1209 11:47:23.262574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"bb1e0efe689536d7366a6ebf4432baa71275c63ede85743aa9403025f57ca310"} Dec 09 11:47:23 crc kubenswrapper[4770]: I1209 11:47:23.262637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8b799" event={"ID":"4002944e-6d2f-4f2b-8d01-e664da3116e2","Type":"ContainerStarted","Data":"104a8ab670484a9fcf66c177484e38de7fca788143ddd9bec92dfd0ba3b76b4a"} Dec 09 11:47:23 crc kubenswrapper[4770]: I1209 11:47:23.263608 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:23 crc kubenswrapper[4770]: I1209 11:47:23.294421 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-8b799" podStartSLOduration=7.443362496 podStartE2EDuration="19.294386562s" podCreationTimestamp="2025-12-09 11:47:04 +0000 UTC" firstStartedPulling="2025-12-09 11:47:05.760374417 +0000 UTC m=+951.001132936" lastFinishedPulling="2025-12-09 11:47:17.611398483 +0000 UTC m=+962.852157002" observedRunningTime="2025-12-09 11:47:23.292093334 +0000 UTC m=+968.532851863" watchObservedRunningTime="2025-12-09 11:47:23.294386562 +0000 UTC m=+968.535145081" Dec 09 11:47:25 crc kubenswrapper[4770]: I1209 11:47:25.230590 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:25 crc kubenswrapper[4770]: I1209 11:47:25.268168 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:26 crc kubenswrapper[4770]: I1209 11:47:26.868594 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7kzks" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.290033 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9"] Dec 09 11:47:28 crc kubenswrapper[4770]: E1209 11:47:28.290674 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="extract-content" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.290691 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="extract-content" Dec 09 11:47:28 crc kubenswrapper[4770]: E1209 11:47:28.290706 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="extract-utilities" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.290714 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="extract-utilities" Dec 09 11:47:28 crc kubenswrapper[4770]: E1209 11:47:28.290726 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="registry-server" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.290736 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="registry-server" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.290892 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="40f6fe58-d863-4a6c-8721-d9d721a21810" containerName="registry-server" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.292077 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.294734 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.310610 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9"] Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.371693 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.372019 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.372155 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kccl\" (UniqueName: \"kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.474276 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.474632 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.474728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kccl\" (UniqueName: \"kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.474929 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.474978 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.498095 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kccl\" (UniqueName: \"kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:28 crc kubenswrapper[4770]: I1209 11:47:28.610658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:29 crc kubenswrapper[4770]: I1209 11:47:29.068736 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9"] Dec 09 11:47:29 crc kubenswrapper[4770]: W1209 11:47:29.076437 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00e48fd8_562c_4212_82c4_40101961789b.slice/crio-951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc WatchSource:0}: Error finding container 951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc: Status 404 returned error can't find the container with id 951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc Dec 09 11:47:29 crc kubenswrapper[4770]: I1209 11:47:29.314459 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerStarted","Data":"2daa503cab47da254e3e5175d7b32c9a690e39bf4dd50a5073a8d4b2bf3c59c3"} Dec 09 11:47:29 crc kubenswrapper[4770]: I1209 11:47:29.314536 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerStarted","Data":"951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc"} Dec 09 11:47:30 crc kubenswrapper[4770]: I1209 11:47:30.323003 4770 generic.go:334] "Generic (PLEG): container finished" podID="00e48fd8-562c-4212-82c4-40101961789b" containerID="2daa503cab47da254e3e5175d7b32c9a690e39bf4dd50a5073a8d4b2bf3c59c3" exitCode=0 Dec 09 11:47:30 crc kubenswrapper[4770]: I1209 11:47:30.323111 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerDied","Data":"2daa503cab47da254e3e5175d7b32c9a690e39bf4dd50a5073a8d4b2bf3c59c3"} Dec 09 11:47:32 crc kubenswrapper[4770]: I1209 11:47:32.473940 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:47:32 crc kubenswrapper[4770]: I1209 11:47:32.474024 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:47:34 crc kubenswrapper[4770]: I1209 11:47:34.377715 4770 generic.go:334] "Generic (PLEG): container finished" podID="00e48fd8-562c-4212-82c4-40101961789b" containerID="02b7931a54edac5a9421896e5cd96bd71db710eaf8988d1d508ff4eae2c92f9b" exitCode=0 Dec 09 11:47:34 crc kubenswrapper[4770]: I1209 11:47:34.377814 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerDied","Data":"02b7931a54edac5a9421896e5cd96bd71db710eaf8988d1d508ff4eae2c92f9b"} Dec 09 11:47:35 crc kubenswrapper[4770]: I1209 11:47:35.233727 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-8b799" Dec 09 11:47:35 crc kubenswrapper[4770]: I1209 11:47:35.385714 4770 generic.go:334] "Generic (PLEG): container finished" podID="00e48fd8-562c-4212-82c4-40101961789b" containerID="92b03dac404b374f0f93a7cb73b0775c945005a00b5b3e35777dc493f63dcc9f" exitCode=0 Dec 09 11:47:35 crc kubenswrapper[4770]: I1209 11:47:35.385772 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerDied","Data":"92b03dac404b374f0f93a7cb73b0775c945005a00b5b3e35777dc493f63dcc9f"} Dec 09 11:47:35 crc kubenswrapper[4770]: I1209 11:47:35.533387 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlj84" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.660111 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.697433 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util\") pod \"00e48fd8-562c-4212-82c4-40101961789b\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.697612 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kccl\" (UniqueName: \"kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl\") pod \"00e48fd8-562c-4212-82c4-40101961789b\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.697675 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle\") pod \"00e48fd8-562c-4212-82c4-40101961789b\" (UID: \"00e48fd8-562c-4212-82c4-40101961789b\") " Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.699012 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle" (OuterVolumeSpecName: "bundle") pod "00e48fd8-562c-4212-82c4-40101961789b" (UID: "00e48fd8-562c-4212-82c4-40101961789b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.703694 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl" (OuterVolumeSpecName: "kube-api-access-6kccl") pod "00e48fd8-562c-4212-82c4-40101961789b" (UID: "00e48fd8-562c-4212-82c4-40101961789b"). InnerVolumeSpecName "kube-api-access-6kccl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.711690 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util" (OuterVolumeSpecName: "util") pod "00e48fd8-562c-4212-82c4-40101961789b" (UID: "00e48fd8-562c-4212-82c4-40101961789b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.799773 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kccl\" (UniqueName: \"kubernetes.io/projected/00e48fd8-562c-4212-82c4-40101961789b-kube-api-access-6kccl\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.799837 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:36 crc kubenswrapper[4770]: I1209 11:47:36.799850 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00e48fd8-562c-4212-82c4-40101961789b-util\") on node \"crc\" DevicePath \"\"" Dec 09 11:47:37 crc kubenswrapper[4770]: I1209 11:47:37.403196 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" event={"ID":"00e48fd8-562c-4212-82c4-40101961789b","Type":"ContainerDied","Data":"951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc"} Dec 09 11:47:37 crc kubenswrapper[4770]: I1209 11:47:37.403251 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="951b4726533cb4ad5aa1162c50a92ce87bd9401adab4e0ef84fccb82bf26d8cc" Dec 09 11:47:37 crc kubenswrapper[4770]: I1209 11:47:37.403303 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.429489 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj"] Dec 09 11:47:41 crc kubenswrapper[4770]: E1209 11:47:41.430374 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="extract" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.430393 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="extract" Dec 09 11:47:41 crc kubenswrapper[4770]: E1209 11:47:41.430406 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="pull" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.430416 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="pull" Dec 09 11:47:41 crc kubenswrapper[4770]: E1209 11:47:41.430428 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="util" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.430436 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="util" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.430591 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e48fd8-562c-4212-82c4-40101961789b" containerName="extract" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.431953 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.434693 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.435028 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-5qd7z" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.446579 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.473292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/68f2995f-9a1d-407a-a602-5a3c91113984-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.473428 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd7f9\" (UniqueName: \"kubernetes.io/projected/68f2995f-9a1d-407a-a602-5a3c91113984-kube-api-access-wd7f9\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.490076 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj"] Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.574251 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/68f2995f-9a1d-407a-a602-5a3c91113984-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.574368 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd7f9\" (UniqueName: \"kubernetes.io/projected/68f2995f-9a1d-407a-a602-5a3c91113984-kube-api-access-wd7f9\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.575172 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/68f2995f-9a1d-407a-a602-5a3c91113984-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.596496 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd7f9\" (UniqueName: \"kubernetes.io/projected/68f2995f-9a1d-407a-a602-5a3c91113984-kube-api-access-wd7f9\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fjqcj\" (UID: \"68f2995f-9a1d-407a-a602-5a3c91113984\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:41 crc kubenswrapper[4770]: I1209 11:47:41.747299 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" Dec 09 11:47:42 crc kubenswrapper[4770]: I1209 11:47:42.310333 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj"] Dec 09 11:47:42 crc kubenswrapper[4770]: I1209 11:47:42.445659 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" event={"ID":"68f2995f-9a1d-407a-a602-5a3c91113984","Type":"ContainerStarted","Data":"90bc28cdcdd09733ed3fdc7e90f5d95aef9867783b7697b0021a7a7284d11ef0"} Dec 09 11:47:53 crc kubenswrapper[4770]: I1209 11:47:53.663768 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" event={"ID":"68f2995f-9a1d-407a-a602-5a3c91113984","Type":"ContainerStarted","Data":"b45af738ab7ad9cb65356b0937b0137b5bee0b9b5458c6fad050d028abd39a3c"} Dec 09 11:47:53 crc kubenswrapper[4770]: I1209 11:47:53.686787 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fjqcj" podStartSLOduration=1.8537762039999999 podStartE2EDuration="12.686766341s" podCreationTimestamp="2025-12-09 11:47:41 +0000 UTC" firstStartedPulling="2025-12-09 11:47:42.323656877 +0000 UTC m=+987.564415396" lastFinishedPulling="2025-12-09 11:47:53.156647014 +0000 UTC m=+998.397405533" observedRunningTime="2025-12-09 11:47:53.684001641 +0000 UTC m=+998.924760180" watchObservedRunningTime="2025-12-09 11:47:53.686766341 +0000 UTC m=+998.927524860" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.774122 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-w75jb"] Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.775403 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.779079 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.780232 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.787785 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-w75jb"] Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.814371 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvvr9\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-kube-api-access-fvvr9\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.814453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.916428 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.917002 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvvr9\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-kube-api-access-fvvr9\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.940970 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:56 crc kubenswrapper[4770]: I1209 11:47:56.941250 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvvr9\" (UniqueName: \"kubernetes.io/projected/07764a7a-7c6b-4839-a163-695ce4b71961-kube-api-access-fvvr9\") pod \"cert-manager-webhook-f4fb5df64-w75jb\" (UID: \"07764a7a-7c6b-4839-a163-695ce4b71961\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:57 crc kubenswrapper[4770]: I1209 11:47:57.101784 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:47:57 crc kubenswrapper[4770]: I1209 11:47:57.403865 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-w75jb"] Dec 09 11:47:57 crc kubenswrapper[4770]: I1209 11:47:57.691822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" event={"ID":"07764a7a-7c6b-4839-a163-695ce4b71961","Type":"ContainerStarted","Data":"15cccfdec7b4a3e8887096be346d6a4f8c7a4378ed182912a33f8689f94aaca5"} Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.006422 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7"] Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.011844 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.020168 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-t4nnk" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.022613 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7"] Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.087508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.088052 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4h4\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-kube-api-access-wz4h4\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.190780 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.190855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4h4\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-kube-api-access-wz4h4\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.219811 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.220000 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4h4\" (UniqueName: \"kubernetes.io/projected/9ae62b97-709d-4d61-8067-808e46f66cc0-kube-api-access-wz4h4\") pod \"cert-manager-cainjector-855d9ccff4-ksfw7\" (UID: \"9ae62b97-709d-4d61-8067-808e46f66cc0\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.351542 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.618793 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7"] Dec 09 11:47:59 crc kubenswrapper[4770]: I1209 11:47:59.728824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" event={"ID":"9ae62b97-709d-4d61-8067-808e46f66cc0","Type":"ContainerStarted","Data":"8d12117b3ef16eb714e680da65ead853429a37c47c99f1c000a69267f9328b02"} Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.474089 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.474848 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.474943 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.476065 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.476183 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24" gracePeriod=600 Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.762426 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24" exitCode=0 Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.762501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24"} Dec 09 11:48:02 crc kubenswrapper[4770]: I1209 11:48:02.762575 4770 scope.go:117] "RemoveContainer" containerID="85a29bdd5f1fa947e4fbc8282761d6e6f256a36dc94daa3cfd8772f5af184e6b" Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.852350 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zxvr2"] Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.853856 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.865337 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-f959g" Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.873772 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zxvr2"] Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.906994 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-bound-sa-token\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:10 crc kubenswrapper[4770]: I1209 11:48:10.907125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tj2l\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-kube-api-access-5tj2l\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:11 crc kubenswrapper[4770]: I1209 11:48:11.010397 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-bound-sa-token\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:11 crc kubenswrapper[4770]: I1209 11:48:11.010482 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tj2l\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-kube-api-access-5tj2l\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:11 crc kubenswrapper[4770]: I1209 11:48:11.034238 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-bound-sa-token\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:11 crc kubenswrapper[4770]: I1209 11:48:11.039450 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tj2l\" (UniqueName: \"kubernetes.io/projected/b28fbeaa-16cf-4e85-bc3d-318554a2c422-kube-api-access-5tj2l\") pod \"cert-manager-86cb77c54b-zxvr2\" (UID: \"b28fbeaa-16cf-4e85-bc3d-318554a2c422\") " pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:11 crc kubenswrapper[4770]: I1209 11:48:11.185343 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-zxvr2" Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:11.982421 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" event={"ID":"9ae62b97-709d-4d61-8067-808e46f66cc0","Type":"ContainerStarted","Data":"da744215621c9ea5edfdeb064f27e3995ce7d226df1bb7c00eb94762e0edd73c"} Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:11.992784 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c"} Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:11.997061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" event={"ID":"07764a7a-7c6b-4839-a163-695ce4b71961","Type":"ContainerStarted","Data":"67ca4b2c8574371b7f5be8ee42dae34842c8d11221bcbc50a08747bdbef0b4c7"} Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:11.998084 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:12.380818 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ksfw7" podStartSLOduration=2.6898374609999998 podStartE2EDuration="14.38079108s" podCreationTimestamp="2025-12-09 11:47:58 +0000 UTC" firstStartedPulling="2025-12-09 11:47:59.641670743 +0000 UTC m=+1004.882429262" lastFinishedPulling="2025-12-09 11:48:11.332624362 +0000 UTC m=+1016.573382881" observedRunningTime="2025-12-09 11:48:12.377119837 +0000 UTC m=+1017.617878356" watchObservedRunningTime="2025-12-09 11:48:12.38079108 +0000 UTC m=+1017.621549599" Dec 09 11:48:12 crc kubenswrapper[4770]: W1209 11:48:12.394509 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb28fbeaa_16cf_4e85_bc3d_318554a2c422.slice/crio-a00b51d6188fe353047d5c86216d904e35089438534624a46bebc9a42f219e07 WatchSource:0}: Error finding container a00b51d6188fe353047d5c86216d904e35089438534624a46bebc9a42f219e07: Status 404 returned error can't find the container with id a00b51d6188fe353047d5c86216d904e35089438534624a46bebc9a42f219e07 Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:12.413530 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-zxvr2"] Dec 09 11:48:12 crc kubenswrapper[4770]: I1209 11:48:12.422298 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" podStartSLOduration=2.54488846 podStartE2EDuration="16.422266865s" podCreationTimestamp="2025-12-09 11:47:56 +0000 UTC" firstStartedPulling="2025-12-09 11:47:57.414157723 +0000 UTC m=+1002.654916242" lastFinishedPulling="2025-12-09 11:48:11.291536128 +0000 UTC m=+1016.532294647" observedRunningTime="2025-12-09 11:48:12.41147285 +0000 UTC m=+1017.652231359" watchObservedRunningTime="2025-12-09 11:48:12.422266865 +0000 UTC m=+1017.663025384" Dec 09 11:48:13 crc kubenswrapper[4770]: I1209 11:48:13.019057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-zxvr2" event={"ID":"b28fbeaa-16cf-4e85-bc3d-318554a2c422","Type":"ContainerStarted","Data":"6040262adbfad3af2cbdb7cdab80973b3ed6cd3259046f0ee7405aa269494323"} Dec 09 11:48:13 crc kubenswrapper[4770]: I1209 11:48:13.019941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-zxvr2" event={"ID":"b28fbeaa-16cf-4e85-bc3d-318554a2c422","Type":"ContainerStarted","Data":"a00b51d6188fe353047d5c86216d904e35089438534624a46bebc9a42f219e07"} Dec 09 11:48:13 crc kubenswrapper[4770]: I1209 11:48:13.078127 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-zxvr2" podStartSLOduration=3.078095808 podStartE2EDuration="3.078095808s" podCreationTimestamp="2025-12-09 11:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:48:13.068127285 +0000 UTC m=+1018.308885804" watchObservedRunningTime="2025-12-09 11:48:13.078095808 +0000 UTC m=+1018.318854327" Dec 09 11:48:17 crc kubenswrapper[4770]: I1209 11:48:17.107836 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-w75jb" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.660429 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.662657 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.666030 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.666139 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.667881 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vffmm" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.688039 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.771814 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7cqr\" (UniqueName: \"kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr\") pod \"openstack-operator-index-l42rc\" (UID: \"d1ade3d0-900e-4354-a003-44b90a8b2ba4\") " pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.873612 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7cqr\" (UniqueName: \"kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr\") pod \"openstack-operator-index-l42rc\" (UID: \"d1ade3d0-900e-4354-a003-44b90a8b2ba4\") " pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.902975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7cqr\" (UniqueName: \"kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr\") pod \"openstack-operator-index-l42rc\" (UID: \"d1ade3d0-900e-4354-a003-44b90a8b2ba4\") " pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:20 crc kubenswrapper[4770]: I1209 11:48:20.989537 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:21 crc kubenswrapper[4770]: I1209 11:48:21.221031 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:22 crc kubenswrapper[4770]: I1209 11:48:22.082317 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l42rc" event={"ID":"d1ade3d0-900e-4354-a003-44b90a8b2ba4","Type":"ContainerStarted","Data":"09cb7d87797386b19430ff06bb5487ddf0b91a4880f1e46341f9c7d248eba066"} Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.018492 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.637980 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-c4v65"] Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.639343 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.643959 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c4v65"] Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.733569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plh68\" (UniqueName: \"kubernetes.io/projected/15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a-kube-api-access-plh68\") pod \"openstack-operator-index-c4v65\" (UID: \"15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a\") " pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.834834 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plh68\" (UniqueName: \"kubernetes.io/projected/15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a-kube-api-access-plh68\") pod \"openstack-operator-index-c4v65\" (UID: \"15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a\") " pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.861878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plh68\" (UniqueName: \"kubernetes.io/projected/15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a-kube-api-access-plh68\") pod \"openstack-operator-index-c4v65\" (UID: \"15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a\") " pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:24 crc kubenswrapper[4770]: I1209 11:48:24.976255 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.100644 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l42rc" event={"ID":"d1ade3d0-900e-4354-a003-44b90a8b2ba4","Type":"ContainerStarted","Data":"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be"} Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.100744 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-l42rc" podUID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" containerName="registry-server" containerID="cri-o://eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be" gracePeriod=2 Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.123257 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-l42rc" podStartSLOduration=1.651314904 podStartE2EDuration="5.123238961s" podCreationTimestamp="2025-12-09 11:48:20 +0000 UTC" firstStartedPulling="2025-12-09 11:48:21.24792442 +0000 UTC m=+1026.488682939" lastFinishedPulling="2025-12-09 11:48:24.719848477 +0000 UTC m=+1029.960606996" observedRunningTime="2025-12-09 11:48:25.113233827 +0000 UTC m=+1030.353992346" watchObservedRunningTime="2025-12-09 11:48:25.123238961 +0000 UTC m=+1030.363997480" Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.846921 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.882423 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c4v65"] Dec 09 11:48:25 crc kubenswrapper[4770]: W1209 11:48:25.883362 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15ff9cb7_d6ab_43d9_bddf_4efa94d8ad7a.slice/crio-fb50db4f0ee70c8acb9e9e4df52dcc976378423b1e95b3dfb6ebbe368e0196ab WatchSource:0}: Error finding container fb50db4f0ee70c8acb9e9e4df52dcc976378423b1e95b3dfb6ebbe368e0196ab: Status 404 returned error can't find the container with id fb50db4f0ee70c8acb9e9e4df52dcc976378423b1e95b3dfb6ebbe368e0196ab Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.940352 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7cqr\" (UniqueName: \"kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr\") pod \"d1ade3d0-900e-4354-a003-44b90a8b2ba4\" (UID: \"d1ade3d0-900e-4354-a003-44b90a8b2ba4\") " Dec 09 11:48:25 crc kubenswrapper[4770]: I1209 11:48:25.949209 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr" (OuterVolumeSpecName: "kube-api-access-w7cqr") pod "d1ade3d0-900e-4354-a003-44b90a8b2ba4" (UID: "d1ade3d0-900e-4354-a003-44b90a8b2ba4"). InnerVolumeSpecName "kube-api-access-w7cqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.042621 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7cqr\" (UniqueName: \"kubernetes.io/projected/d1ade3d0-900e-4354-a003-44b90a8b2ba4-kube-api-access-w7cqr\") on node \"crc\" DevicePath \"\"" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.108600 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c4v65" event={"ID":"15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a","Type":"ContainerStarted","Data":"fb50db4f0ee70c8acb9e9e4df52dcc976378423b1e95b3dfb6ebbe368e0196ab"} Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.109972 4770 generic.go:334] "Generic (PLEG): container finished" podID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" containerID="eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be" exitCode=0 Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.110010 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l42rc" event={"ID":"d1ade3d0-900e-4354-a003-44b90a8b2ba4","Type":"ContainerDied","Data":"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be"} Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.110032 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l42rc" event={"ID":"d1ade3d0-900e-4354-a003-44b90a8b2ba4","Type":"ContainerDied","Data":"09cb7d87797386b19430ff06bb5487ddf0b91a4880f1e46341f9c7d248eba066"} Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.110049 4770 scope.go:117] "RemoveContainer" containerID="eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.110248 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l42rc" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.140627 4770 scope.go:117] "RemoveContainer" containerID="eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be" Dec 09 11:48:26 crc kubenswrapper[4770]: E1209 11:48:26.141995 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be\": container with ID starting with eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be not found: ID does not exist" containerID="eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.142054 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be"} err="failed to get container status \"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be\": rpc error: code = NotFound desc = could not find container \"eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be\": container with ID starting with eb4bc59e5c22c3531f9669ed2d1846227f1ddb3252d0a76eeb9a31a89373f7be not found: ID does not exist" Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.146778 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:26 crc kubenswrapper[4770]: I1209 11:48:26.152596 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-l42rc"] Dec 09 11:48:27 crc kubenswrapper[4770]: I1209 11:48:27.119917 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c4v65" event={"ID":"15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a","Type":"ContainerStarted","Data":"555a5d17e48456b49574839f2ce6e010ae939e311852453de3c6ab68fb0c57a8"} Dec 09 11:48:27 crc kubenswrapper[4770]: I1209 11:48:27.350887 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" path="/var/lib/kubelet/pods/d1ade3d0-900e-4354-a003-44b90a8b2ba4/volumes" Dec 09 11:48:34 crc kubenswrapper[4770]: I1209 11:48:34.977189 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:34 crc kubenswrapper[4770]: I1209 11:48:34.977835 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:35 crc kubenswrapper[4770]: I1209 11:48:35.025566 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:35 crc kubenswrapper[4770]: I1209 11:48:35.039564 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-c4v65" podStartSLOduration=10.615997956 podStartE2EDuration="11.039545274s" podCreationTimestamp="2025-12-09 11:48:24 +0000 UTC" firstStartedPulling="2025-12-09 11:48:25.889535763 +0000 UTC m=+1031.130294272" lastFinishedPulling="2025-12-09 11:48:26.313083071 +0000 UTC m=+1031.553841590" observedRunningTime="2025-12-09 11:48:27.139139472 +0000 UTC m=+1032.379898001" watchObservedRunningTime="2025-12-09 11:48:35.039545274 +0000 UTC m=+1040.280303793" Dec 09 11:48:35 crc kubenswrapper[4770]: I1209 11:48:35.198789 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-c4v65" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.046474 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx"] Dec 09 11:48:42 crc kubenswrapper[4770]: E1209 11:48:42.047847 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" containerName="registry-server" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.047868 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" containerName="registry-server" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.048086 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ade3d0-900e-4354-a003-44b90a8b2ba4" containerName="registry-server" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.049397 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.052216 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-tz2gp" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.058013 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx"] Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.195074 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.195225 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlvjr\" (UniqueName: \"kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.195348 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.296568 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.296627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlvjr\" (UniqueName: \"kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.296669 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.297079 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.297096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.316259 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlvjr\" (UniqueName: \"kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.371792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:42 crc kubenswrapper[4770]: I1209 11:48:42.781040 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx"] Dec 09 11:48:43 crc kubenswrapper[4770]: I1209 11:48:43.235057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" event={"ID":"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa","Type":"ContainerStarted","Data":"2e559bc91f067d347e8d934fbd328b170eae65c2826abccbbe702ce8351dcfb9"} Dec 09 11:48:45 crc kubenswrapper[4770]: I1209 11:48:45.254989 4770 generic.go:334] "Generic (PLEG): container finished" podID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerID="c59353f44ec92616147ccec18f25bb8526a442a046e5fdceb09da694d22ddd67" exitCode=0 Dec 09 11:48:45 crc kubenswrapper[4770]: I1209 11:48:45.255054 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" event={"ID":"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa","Type":"ContainerDied","Data":"c59353f44ec92616147ccec18f25bb8526a442a046e5fdceb09da694d22ddd67"} Dec 09 11:48:47 crc kubenswrapper[4770]: I1209 11:48:47.272025 4770 generic.go:334] "Generic (PLEG): container finished" podID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerID="e8b6c1fb5c32b1bedfb3ae8dd5d6997756bb38e688c61c04ba4fdd21a5500c1f" exitCode=0 Dec 09 11:48:47 crc kubenswrapper[4770]: I1209 11:48:47.272166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" event={"ID":"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa","Type":"ContainerDied","Data":"e8b6c1fb5c32b1bedfb3ae8dd5d6997756bb38e688c61c04ba4fdd21a5500c1f"} Dec 09 11:48:48 crc kubenswrapper[4770]: I1209 11:48:48.283614 4770 generic.go:334] "Generic (PLEG): container finished" podID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerID="6ff45c2be27063a2b0c215ad40d2643db86129713b7ffca8947d855c7bc3464f" exitCode=0 Dec 09 11:48:48 crc kubenswrapper[4770]: I1209 11:48:48.283716 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" event={"ID":"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa","Type":"ContainerDied","Data":"6ff45c2be27063a2b0c215ad40d2643db86129713b7ffca8947d855c7bc3464f"} Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.575291 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.790350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlvjr\" (UniqueName: \"kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr\") pod \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.790414 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle\") pod \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.790550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util\") pod \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\" (UID: \"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa\") " Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.791319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle" (OuterVolumeSpecName: "bundle") pod "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" (UID: "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.798172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr" (OuterVolumeSpecName: "kube-api-access-wlvjr") pod "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" (UID: "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa"). InnerVolumeSpecName "kube-api-access-wlvjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.806371 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util" (OuterVolumeSpecName: "util") pod "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" (UID: "c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.892445 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlvjr\" (UniqueName: \"kubernetes.io/projected/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-kube-api-access-wlvjr\") on node \"crc\" DevicePath \"\"" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.892528 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:48:49 crc kubenswrapper[4770]: I1209 11:48:49.892542 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa-util\") on node \"crc\" DevicePath \"\"" Dec 09 11:48:50 crc kubenswrapper[4770]: I1209 11:48:50.304083 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" event={"ID":"c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa","Type":"ContainerDied","Data":"2e559bc91f067d347e8d934fbd328b170eae65c2826abccbbe702ce8351dcfb9"} Dec 09 11:48:50 crc kubenswrapper[4770]: I1209 11:48:50.304140 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e559bc91f067d347e8d934fbd328b170eae65c2826abccbbe702ce8351dcfb9" Dec 09 11:48:50 crc kubenswrapper[4770]: I1209 11:48:50.304207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.717796 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5"] Dec 09 11:48:54 crc kubenswrapper[4770]: E1209 11:48:54.719817 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="extract" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.719938 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="extract" Dec 09 11:48:54 crc kubenswrapper[4770]: E1209 11:48:54.720070 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="util" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.720145 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="util" Dec 09 11:48:54 crc kubenswrapper[4770]: E1209 11:48:54.720204 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="pull" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.720260 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="pull" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.720473 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa" containerName="extract" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.721219 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.731377 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-nswsd" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.757206 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5"] Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.777233 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h8kg\" (UniqueName: \"kubernetes.io/projected/ddbcaf12-235d-4b39-ba48-35c3fb15b5a8-kube-api-access-4h8kg\") pod \"openstack-operator-controller-operator-7979d445b4-qvrn5\" (UID: \"ddbcaf12-235d-4b39-ba48-35c3fb15b5a8\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.879260 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h8kg\" (UniqueName: \"kubernetes.io/projected/ddbcaf12-235d-4b39-ba48-35c3fb15b5a8-kube-api-access-4h8kg\") pod \"openstack-operator-controller-operator-7979d445b4-qvrn5\" (UID: \"ddbcaf12-235d-4b39-ba48-35c3fb15b5a8\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:48:54 crc kubenswrapper[4770]: I1209 11:48:54.908810 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h8kg\" (UniqueName: \"kubernetes.io/projected/ddbcaf12-235d-4b39-ba48-35c3fb15b5a8-kube-api-access-4h8kg\") pod \"openstack-operator-controller-operator-7979d445b4-qvrn5\" (UID: \"ddbcaf12-235d-4b39-ba48-35c3fb15b5a8\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:48:55 crc kubenswrapper[4770]: I1209 11:48:55.045978 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:48:55 crc kubenswrapper[4770]: I1209 11:48:55.369119 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5"] Dec 09 11:48:55 crc kubenswrapper[4770]: W1209 11:48:55.372457 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddbcaf12_235d_4b39_ba48_35c3fb15b5a8.slice/crio-203fd845b2e74d656cc746a155e7ea5627a55585a7304dbfd92f4055e02066e6 WatchSource:0}: Error finding container 203fd845b2e74d656cc746a155e7ea5627a55585a7304dbfd92f4055e02066e6: Status 404 returned error can't find the container with id 203fd845b2e74d656cc746a155e7ea5627a55585a7304dbfd92f4055e02066e6 Dec 09 11:48:56 crc kubenswrapper[4770]: I1209 11:48:56.360341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" event={"ID":"ddbcaf12-235d-4b39-ba48-35c3fb15b5a8","Type":"ContainerStarted","Data":"203fd845b2e74d656cc746a155e7ea5627a55585a7304dbfd92f4055e02066e6"} Dec 09 11:49:02 crc kubenswrapper[4770]: I1209 11:49:02.482464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" event={"ID":"ddbcaf12-235d-4b39-ba48-35c3fb15b5a8","Type":"ContainerStarted","Data":"3467e7f89c63eeef3a6a017f53e02d1305689cd3e2d57540535c2dc878346f02"} Dec 09 11:49:02 crc kubenswrapper[4770]: I1209 11:49:02.483490 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:49:02 crc kubenswrapper[4770]: I1209 11:49:02.521297 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" podStartSLOduration=1.629211811 podStartE2EDuration="8.521257327s" podCreationTimestamp="2025-12-09 11:48:54 +0000 UTC" firstStartedPulling="2025-12-09 11:48:55.375934141 +0000 UTC m=+1060.616692660" lastFinishedPulling="2025-12-09 11:49:02.267979657 +0000 UTC m=+1067.508738176" observedRunningTime="2025-12-09 11:49:02.517143272 +0000 UTC m=+1067.757901801" watchObservedRunningTime="2025-12-09 11:49:02.521257327 +0000 UTC m=+1067.762015846" Dec 09 11:49:15 crc kubenswrapper[4770]: I1209 11:49:15.050595 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-qvrn5" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.681377 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.685893 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.688672 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-nz4rd" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.704975 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.720432 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.722419 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.726358 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-sqtf9" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.764552 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.769517 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89l8h\" (UniqueName: \"kubernetes.io/projected/303a0c89-f011-4953-ac5c-33929697ccd1-kube-api-access-89l8h\") pod \"barbican-operator-controller-manager-7d9dfd778-h5w2l\" (UID: \"303a0c89-f011-4953-ac5c-33929697ccd1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.774130 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.777654 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.778717 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4ggg\" (UniqueName: \"kubernetes.io/projected/cf58af47-dce9-43df-90cc-7c642e3b2f3b-kube-api-access-d4ggg\") pod \"cinder-operator-controller-manager-6c677c69b-5gj7z\" (UID: \"cf58af47-dce9-43df-90cc-7c642e3b2f3b\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.780196 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.786878 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-jcfqs" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.806757 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.808096 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.811573 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jbvsg" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.827551 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.834552 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.841930 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qzwxx" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.842802 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.879568 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.880784 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m74xq\" (UniqueName: \"kubernetes.io/projected/3588fb02-b477-40dd-976e-f9b8e0508324-kube-api-access-m74xq\") pod \"glance-operator-controller-manager-5697bb5779-p69k8\" (UID: \"3588fb02-b477-40dd-976e-f9b8e0508324\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.880912 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89l8h\" (UniqueName: \"kubernetes.io/projected/303a0c89-f011-4953-ac5c-33929697ccd1-kube-api-access-89l8h\") pod \"barbican-operator-controller-manager-7d9dfd778-h5w2l\" (UID: \"303a0c89-f011-4953-ac5c-33929697ccd1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.880966 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4ggg\" (UniqueName: \"kubernetes.io/projected/cf58af47-dce9-43df-90cc-7c642e3b2f3b-kube-api-access-d4ggg\") pod \"cinder-operator-controller-manager-6c677c69b-5gj7z\" (UID: \"cf58af47-dce9-43df-90cc-7c642e3b2f3b\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.881038 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jkc2\" (UniqueName: \"kubernetes.io/projected/9a0fffd2-c599-4a15-b1f0-3c404d44a7bc-kube-api-access-9jkc2\") pod \"designate-operator-controller-manager-697fb699cf-r89wl\" (UID: \"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.881076 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ss9n\" (UniqueName: \"kubernetes.io/projected/2537825c-aa47-40ac-bb98-d9793449e9dd-kube-api-access-6ss9n\") pod \"heat-operator-controller-manager-5f64f6f8bb-56gd2\" (UID: \"2537825c-aa47-40ac-bb98-d9793449e9dd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.895841 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.897314 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.906581 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8ppkp" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.926213 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4ggg\" (UniqueName: \"kubernetes.io/projected/cf58af47-dce9-43df-90cc-7c642e3b2f3b-kube-api-access-d4ggg\") pod \"cinder-operator-controller-manager-6c677c69b-5gj7z\" (UID: \"cf58af47-dce9-43df-90cc-7c642e3b2f3b\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.940021 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.941630 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.943437 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89l8h\" (UniqueName: \"kubernetes.io/projected/303a0c89-f011-4953-ac5c-33929697ccd1-kube-api-access-89l8h\") pod \"barbican-operator-controller-manager-7d9dfd778-h5w2l\" (UID: \"303a0c89-f011-4953-ac5c-33929697ccd1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.949420 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r6nxf" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.950204 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.975603 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9"] Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.984730 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.984799 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cjct\" (UniqueName: \"kubernetes.io/projected/d241ba8f-5573-466a-88f2-89ec9f14cc29-kube-api-access-8cjct\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.984913 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jkc2\" (UniqueName: \"kubernetes.io/projected/9a0fffd2-c599-4a15-b1f0-3c404d44a7bc-kube-api-access-9jkc2\") pod \"designate-operator-controller-manager-697fb699cf-r89wl\" (UID: \"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.984945 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwznj\" (UniqueName: \"kubernetes.io/projected/23265ff6-b252-4fc0-a3b4-6b59deea7c69-kube-api-access-fwznj\") pod \"horizon-operator-controller-manager-68c6d99b8f-4lvc9\" (UID: \"23265ff6-b252-4fc0-a3b4-6b59deea7c69\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.984980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ss9n\" (UniqueName: \"kubernetes.io/projected/2537825c-aa47-40ac-bb98-d9793449e9dd-kube-api-access-6ss9n\") pod \"heat-operator-controller-manager-5f64f6f8bb-56gd2\" (UID: \"2537825c-aa47-40ac-bb98-d9793449e9dd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:49:42 crc kubenswrapper[4770]: I1209 11:49:42.985038 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m74xq\" (UniqueName: \"kubernetes.io/projected/3588fb02-b477-40dd-976e-f9b8e0508324-kube-api-access-m74xq\") pod \"glance-operator-controller-manager-5697bb5779-p69k8\" (UID: \"3588fb02-b477-40dd-976e-f9b8e0508324\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.006654 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.017631 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.019503 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.038356 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ss9n\" (UniqueName: \"kubernetes.io/projected/2537825c-aa47-40ac-bb98-d9793449e9dd-kube-api-access-6ss9n\") pod \"heat-operator-controller-manager-5f64f6f8bb-56gd2\" (UID: \"2537825c-aa47-40ac-bb98-d9793449e9dd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.040948 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.045950 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jkc2\" (UniqueName: \"kubernetes.io/projected/9a0fffd2-c599-4a15-b1f0-3c404d44a7bc-kube-api-access-9jkc2\") pod \"designate-operator-controller-manager-697fb699cf-r89wl\" (UID: \"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.048078 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.049999 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.054077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m74xq\" (UniqueName: \"kubernetes.io/projected/3588fb02-b477-40dd-976e-f9b8e0508324-kube-api-access-m74xq\") pod \"glance-operator-controller-manager-5697bb5779-p69k8\" (UID: \"3588fb02-b477-40dd-976e-f9b8e0508324\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.055246 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-vt6z5" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.056793 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-jnm8c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.061814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.118625 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.132520 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwznj\" (UniqueName: \"kubernetes.io/projected/23265ff6-b252-4fc0-a3b4-6b59deea7c69-kube-api-access-fwznj\") pod \"horizon-operator-controller-manager-68c6d99b8f-4lvc9\" (UID: \"23265ff6-b252-4fc0-a3b4-6b59deea7c69\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.132612 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6zh\" (UniqueName: \"kubernetes.io/projected/c8cbcbfa-fe34-4fae-96b6-04a413418bf0-kube-api-access-cg6zh\") pod \"ironic-operator-controller-manager-967d97867-zhc7c\" (UID: \"c8cbcbfa-fe34-4fae-96b6-04a413418bf0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.132693 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.132730 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cjct\" (UniqueName: \"kubernetes.io/projected/d241ba8f-5573-466a-88f2-89ec9f14cc29-kube-api-access-8cjct\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.132774 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxpq2\" (UniqueName: \"kubernetes.io/projected/7a873c67-c1eb-4681-b18e-25e1580bb11c-kube-api-access-lxpq2\") pod \"keystone-operator-controller-manager-7765d96ddf-846ng\" (UID: \"7a873c67-c1eb-4681-b18e-25e1580bb11c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:49:43 crc kubenswrapper[4770]: E1209 11:49:43.133531 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:43 crc kubenswrapper[4770]: E1209 11:49:43.133595 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:43.633574673 +0000 UTC m=+1108.874333192 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.138215 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.172270 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.191413 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwznj\" (UniqueName: \"kubernetes.io/projected/23265ff6-b252-4fc0-a3b4-6b59deea7c69-kube-api-access-fwznj\") pod \"horizon-operator-controller-manager-68c6d99b8f-4lvc9\" (UID: \"23265ff6-b252-4fc0-a3b4-6b59deea7c69\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.194235 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.202569 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cjct\" (UniqueName: \"kubernetes.io/projected/d241ba8f-5573-466a-88f2-89ec9f14cc29-kube-api-access-8cjct\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.234813 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.236705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6zh\" (UniqueName: \"kubernetes.io/projected/c8cbcbfa-fe34-4fae-96b6-04a413418bf0-kube-api-access-cg6zh\") pod \"ironic-operator-controller-manager-967d97867-zhc7c\" (UID: \"c8cbcbfa-fe34-4fae-96b6-04a413418bf0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.236856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxpq2\" (UniqueName: \"kubernetes.io/projected/7a873c67-c1eb-4681-b18e-25e1580bb11c-kube-api-access-lxpq2\") pod \"keystone-operator-controller-manager-7765d96ddf-846ng\" (UID: \"7a873c67-c1eb-4681-b18e-25e1580bb11c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.240044 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.243124 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.289299 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6ckx5" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.290208 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.320446 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.321972 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.335307 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxpq2\" (UniqueName: \"kubernetes.io/projected/7a873c67-c1eb-4681-b18e-25e1580bb11c-kube-api-access-lxpq2\") pod \"keystone-operator-controller-manager-7765d96ddf-846ng\" (UID: \"7a873c67-c1eb-4681-b18e-25e1580bb11c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.339561 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9q5j\" (UniqueName: \"kubernetes.io/projected/211dd097-458c-435c-b3b0-b952dc866fac-kube-api-access-t9q5j\") pod \"manila-operator-controller-manager-5b5fd79c9c-vvf9z\" (UID: \"211dd097-458c-435c-b3b0-b952dc866fac\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.340334 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.340520 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6zh\" (UniqueName: \"kubernetes.io/projected/c8cbcbfa-fe34-4fae-96b6-04a413418bf0-kube-api-access-cg6zh\") pod \"ironic-operator-controller-manager-967d97867-zhc7c\" (UID: \"c8cbcbfa-fe34-4fae-96b6-04a413418bf0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.362141 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vjw4m" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.385237 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.386380 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.394783 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.409784 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wl7lq" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.416155 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.417723 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.429537 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8d4zn" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.446051 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.464551 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl55g\" (UniqueName: \"kubernetes.io/projected/8c0e510c-adb0-43f4-b11d-05b378876a46-kube-api-access-nl55g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jv56l\" (UID: \"8c0e510c-adb0-43f4-b11d-05b378876a46\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.464773 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9q5j\" (UniqueName: \"kubernetes.io/projected/211dd097-458c-435c-b3b0-b952dc866fac-kube-api-access-t9q5j\") pod \"manila-operator-controller-manager-5b5fd79c9c-vvf9z\" (UID: \"211dd097-458c-435c-b3b0-b952dc866fac\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.464813 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgg7f\" (UniqueName: \"kubernetes.io/projected/fd01836c-e495-43e0-8552-e17a41352a3d-kube-api-access-wgg7f\") pod \"mariadb-operator-controller-manager-79c8c4686c-5jr9q\" (UID: \"fd01836c-e495-43e0-8552-e17a41352a3d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.464868 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wljmn"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.466094 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.507798 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.508145 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.511366 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.560004 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wljmn"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.578973 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.592857 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.594435 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.595432 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.627357 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.647444 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xszh4"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.660291 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673082 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673220 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxs8z\" (UniqueName: \"kubernetes.io/projected/520a9036-0b49-4236-90cf-ada8df687ad9-kube-api-access-bxs8z\") pod \"nova-operator-controller-manager-697bc559fc-vjw9d\" (UID: \"520a9036-0b49-4236-90cf-ada8df687ad9\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgg7f\" (UniqueName: \"kubernetes.io/projected/fd01836c-e495-43e0-8552-e17a41352a3d-kube-api-access-wgg7f\") pod \"mariadb-operator-controller-manager-79c8c4686c-5jr9q\" (UID: \"fd01836c-e495-43e0-8552-e17a41352a3d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673426 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl55g\" (UniqueName: \"kubernetes.io/projected/8c0e510c-adb0-43f4-b11d-05b378876a46-kube-api-access-nl55g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jv56l\" (UID: \"8c0e510c-adb0-43f4-b11d-05b378876a46\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673492 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xktb6\" (UniqueName: \"kubernetes.io/projected/7eea33f2-55bf-4542-82db-7cf987f29a63-kube-api-access-xktb6\") pod \"octavia-operator-controller-manager-998648c74-wljmn\" (UID: \"7eea33f2-55bf-4542-82db-7cf987f29a63\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.673686 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd"] Dec 09 11:49:43 crc kubenswrapper[4770]: E1209 11:49:43.673823 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:43 crc kubenswrapper[4770]: E1209 11:49:43.673925 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:44.673878279 +0000 UTC m=+1109.914636808 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.701016 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xszh4"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.726965 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.728601 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.732995 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.734629 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.737009 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.776741 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.778005 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xktb6\" (UniqueName: \"kubernetes.io/projected/7eea33f2-55bf-4542-82db-7cf987f29a63-kube-api-access-xktb6\") pod \"octavia-operator-controller-manager-998648c74-wljmn\" (UID: \"7eea33f2-55bf-4542-82db-7cf987f29a63\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.778491 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trxbp\" (UniqueName: \"kubernetes.io/projected/70ea2809-e1eb-4cf6-bd48-21ad18c031a4-kube-api-access-trxbp\") pod \"placement-operator-controller-manager-78f8948974-xszh4\" (UID: \"70ea2809-e1eb-4cf6-bd48-21ad18c031a4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.778548 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxs8z\" (UniqueName: \"kubernetes.io/projected/520a9036-0b49-4236-90cf-ada8df687ad9-kube-api-access-bxs8z\") pod \"nova-operator-controller-manager-697bc559fc-vjw9d\" (UID: \"520a9036-0b49-4236-90cf-ada8df687ad9\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.778645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljt9l\" (UniqueName: \"kubernetes.io/projected/86d62982-f66a-4184-8e15-9336bfca2b8f-kube-api-access-ljt9l\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.778690 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4g62\" (UniqueName: \"kubernetes.io/projected/764304ee-eb30-434a-b884-f7e455576411-kube-api-access-w4g62\") pod \"ovn-operator-controller-manager-b6456fdb6-w86cd\" (UID: \"764304ee-eb30-434a-b884-f7e455576411\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.789030 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.801803 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bggpm"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.803004 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.821479 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8ds5f" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.821787 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.822007 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fsj6b" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.822293 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jngvm" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.828595 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-h8wqz" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.834485 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-5g9td" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.835926 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hngs2" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.836588 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7655s" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.842855 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxs8z\" (UniqueName: \"kubernetes.io/projected/520a9036-0b49-4236-90cf-ada8df687ad9-kube-api-access-bxs8z\") pod \"nova-operator-controller-manager-697bc559fc-vjw9d\" (UID: \"520a9036-0b49-4236-90cf-ada8df687ad9\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.864471 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bggpm"] Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.869687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xktb6\" (UniqueName: \"kubernetes.io/projected/7eea33f2-55bf-4542-82db-7cf987f29a63-kube-api-access-xktb6\") pod \"octavia-operator-controller-manager-998648c74-wljmn\" (UID: \"7eea33f2-55bf-4542-82db-7cf987f29a63\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:49:43 crc kubenswrapper[4770]: I1209 11:49:43.870712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9q5j\" (UniqueName: \"kubernetes.io/projected/211dd097-458c-435c-b3b0-b952dc866fac-kube-api-access-t9q5j\") pod \"manila-operator-controller-manager-5b5fd79c9c-vvf9z\" (UID: \"211dd097-458c-435c-b3b0-b952dc866fac\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.899212 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqjmz\" (UniqueName: \"kubernetes.io/projected/3da14cb0-e4a9-4fe0-bf66-616a2ed438eb-kube-api-access-gqjmz\") pod \"swift-operator-controller-manager-9d58d64bc-b7x22\" (UID: \"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.899281 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trxbp\" (UniqueName: \"kubernetes.io/projected/70ea2809-e1eb-4cf6-bd48-21ad18c031a4-kube-api-access-trxbp\") pod \"placement-operator-controller-manager-78f8948974-xszh4\" (UID: \"70ea2809-e1eb-4cf6-bd48-21ad18c031a4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.899378 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrwzm\" (UniqueName: \"kubernetes.io/projected/daa9d1b4-0867-4e4b-a5a4-037de4506c62-kube-api-access-zrwzm\") pod \"telemetry-operator-controller-manager-58d5ff84df-96cv7\" (UID: \"daa9d1b4-0867-4e4b-a5a4-037de4506c62\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.899437 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljt9l\" (UniqueName: \"kubernetes.io/projected/86d62982-f66a-4184-8e15-9336bfca2b8f-kube-api-access-ljt9l\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.940164 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgg7f\" (UniqueName: \"kubernetes.io/projected/fd01836c-e495-43e0-8552-e17a41352a3d-kube-api-access-wgg7f\") pod \"mariadb-operator-controller-manager-79c8c4686c-5jr9q\" (UID: \"fd01836c-e495-43e0-8552-e17a41352a3d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.963600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4g62\" (UniqueName: \"kubernetes.io/projected/764304ee-eb30-434a-b884-f7e455576411-kube-api-access-w4g62\") pod \"ovn-operator-controller-manager-b6456fdb6-w86cd\" (UID: \"764304ee-eb30-434a-b884-f7e455576411\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.963874 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:44 crc kubenswrapper[4770]: E1209 11:49:43.964348 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:44 crc kubenswrapper[4770]: E1209 11:49:43.964409 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert podName:86d62982-f66a-4184-8e15-9336bfca2b8f nodeName:}" failed. No retries permitted until 2025-12-09 11:49:44.464389664 +0000 UTC m=+1109.705148183 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" (UID: "86d62982-f66a-4184-8e15-9336bfca2b8f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.965822 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.969523 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl55g\" (UniqueName: \"kubernetes.io/projected/8c0e510c-adb0-43f4-b11d-05b378876a46-kube-api-access-nl55g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jv56l\" (UID: \"8c0e510c-adb0-43f4-b11d-05b378876a46\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:43.983260 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts"] Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.057141 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.060531 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.061431 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7lrfl" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.061851 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.075706 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n49zh\" (UniqueName: \"kubernetes.io/projected/7af1b779-e5fd-40a2-8913-9a334a300efb-kube-api-access-n49zh\") pod \"watcher-operator-controller-manager-667bd8d554-bq5ts\" (UID: \"7af1b779-e5fd-40a2-8913-9a334a300efb\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.075769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqjmz\" (UniqueName: \"kubernetes.io/projected/3da14cb0-e4a9-4fe0-bf66-616a2ed438eb-kube-api-access-gqjmz\") pod \"swift-operator-controller-manager-9d58d64bc-b7x22\" (UID: \"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.075835 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrwzm\" (UniqueName: \"kubernetes.io/projected/daa9d1b4-0867-4e4b-a5a4-037de4506c62-kube-api-access-zrwzm\") pod \"telemetry-operator-controller-manager-58d5ff84df-96cv7\" (UID: \"daa9d1b4-0867-4e4b-a5a4-037de4506c62\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.075944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8zjc\" (UniqueName: \"kubernetes.io/projected/049491b6-ec1f-421a-bfb3-bc7c63f27f8c-kube-api-access-w8zjc\") pod \"test-operator-controller-manager-5854674fcc-bggpm\" (UID: \"049491b6-ec1f-421a-bfb3-bc7c63f27f8c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.088960 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.102764 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljt9l\" (UniqueName: \"kubernetes.io/projected/86d62982-f66a-4184-8e15-9336bfca2b8f-kube-api-access-ljt9l\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.179860 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.187612 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrwzm\" (UniqueName: \"kubernetes.io/projected/daa9d1b4-0867-4e4b-a5a4-037de4506c62-kube-api-access-zrwzm\") pod \"telemetry-operator-controller-manager-58d5ff84df-96cv7\" (UID: \"daa9d1b4-0867-4e4b-a5a4-037de4506c62\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.188344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n49zh\" (UniqueName: \"kubernetes.io/projected/7af1b779-e5fd-40a2-8913-9a334a300efb-kube-api-access-n49zh\") pod \"watcher-operator-controller-manager-667bd8d554-bq5ts\" (UID: \"7af1b779-e5fd-40a2-8913-9a334a300efb\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.188614 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8zjc\" (UniqueName: \"kubernetes.io/projected/049491b6-ec1f-421a-bfb3-bc7c63f27f8c-kube-api-access-w8zjc\") pod \"test-operator-controller-manager-5854674fcc-bggpm\" (UID: \"049491b6-ec1f-421a-bfb3-bc7c63f27f8c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.285706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqjmz\" (UniqueName: \"kubernetes.io/projected/3da14cb0-e4a9-4fe0-bf66-616a2ed438eb-kube-api-access-gqjmz\") pod \"swift-operator-controller-manager-9d58d64bc-b7x22\" (UID: \"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.286558 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trxbp\" (UniqueName: \"kubernetes.io/projected/70ea2809-e1eb-4cf6-bd48-21ad18c031a4-kube-api-access-trxbp\") pod \"placement-operator-controller-manager-78f8948974-xszh4\" (UID: \"70ea2809-e1eb-4cf6-bd48-21ad18c031a4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.359804 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8zjc\" (UniqueName: \"kubernetes.io/projected/049491b6-ec1f-421a-bfb3-bc7c63f27f8c-kube-api-access-w8zjc\") pod \"test-operator-controller-manager-5854674fcc-bggpm\" (UID: \"049491b6-ec1f-421a-bfb3-bc7c63f27f8c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.415317 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.440725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4g62\" (UniqueName: \"kubernetes.io/projected/764304ee-eb30-434a-b884-f7e455576411-kube-api-access-w4g62\") pod \"ovn-operator-controller-manager-b6456fdb6-w86cd\" (UID: \"764304ee-eb30-434a-b884-f7e455576411\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.446550 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.448663 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n49zh\" (UniqueName: \"kubernetes.io/projected/7af1b779-e5fd-40a2-8913-9a334a300efb-kube-api-access-n49zh\") pod \"watcher-operator-controller-manager-667bd8d554-bq5ts\" (UID: \"7af1b779-e5fd-40a2-8913-9a334a300efb\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.459526 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts"] Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.469871 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.474627 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8"] Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.500695 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.520860 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.525036 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.526856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:44 crc kubenswrapper[4770]: E1209 11:49:44.527494 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.527498 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 09 11:49:44 crc kubenswrapper[4770]: E1209 11:49:44.527791 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert podName:86d62982-f66a-4184-8e15-9336bfca2b8f nodeName:}" failed. No retries permitted until 2025-12-09 11:49:45.527575622 +0000 UTC m=+1110.768334141 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" (UID: "86d62982-f66a-4184-8e15-9336bfca2b8f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.530247 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9lgpq" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.530435 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.542271 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:49:44 crc kubenswrapper[4770]: I1209 11:49:44.555390 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8"] Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.455707 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.458287 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.458380 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:47.458348945 +0000 UTC m=+1112.699107534 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.516950 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h"] Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.519383 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.543130 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h"] Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.570942 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-pldqd" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.576351 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.576835 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v24pj\" (UniqueName: \"kubernetes.io/projected/30458aa5-ed86-491e-b2bd-353725caf57d-kube-api-access-v24pj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h6w6h\" (UID: \"30458aa5-ed86-491e-b2bd-353725caf57d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.577013 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.577303 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t76qr\" (UniqueName: \"kubernetes.io/projected/c57a5114-2563-4390-855b-402e73c907d6-kube-api-access-t76qr\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.577524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.579061 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.581798 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert podName:86d62982-f66a-4184-8e15-9336bfca2b8f nodeName:}" failed. No retries permitted until 2025-12-09 11:49:47.581761413 +0000 UTC m=+1112.822519942 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" (UID: "86d62982-f66a-4184-8e15-9336bfca2b8f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.609633 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z"] Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.686083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.686224 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.686273 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v24pj\" (UniqueName: \"kubernetes.io/projected/30458aa5-ed86-491e-b2bd-353725caf57d-kube-api-access-v24pj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h6w6h\" (UID: \"30458aa5-ed86-491e-b2bd-353725caf57d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.686313 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t76qr\" (UniqueName: \"kubernetes.io/projected/c57a5114-2563-4390-855b-402e73c907d6-kube-api-access-t76qr\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.686614 4770 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.686760 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:46.186732142 +0000 UTC m=+1111.427490661 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "metrics-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.686785 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: E1209 11:49:45.688371 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:46.188355023 +0000 UTC m=+1111.429113552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "webhook-server-cert" not found Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.972470 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t76qr\" (UniqueName: \"kubernetes.io/projected/c57a5114-2563-4390-855b-402e73c907d6-kube-api-access-t76qr\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:45 crc kubenswrapper[4770]: I1209 11:49:45.978206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v24pj\" (UniqueName: \"kubernetes.io/projected/30458aa5-ed86-491e-b2bd-353725caf57d-kube-api-access-v24pj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h6w6h\" (UID: \"30458aa5-ed86-491e-b2bd-353725caf57d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" Dec 09 11:49:46 crc kubenswrapper[4770]: I1209 11:49:46.067247 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" Dec 09 11:49:46 crc kubenswrapper[4770]: I1209 11:49:46.261716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:46 crc kubenswrapper[4770]: I1209 11:49:46.261932 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:46 crc kubenswrapper[4770]: E1209 11:49:46.262029 4770 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 11:49:46 crc kubenswrapper[4770]: E1209 11:49:46.262117 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 11:49:46 crc kubenswrapper[4770]: E1209 11:49:46.262170 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:47.26212867 +0000 UTC m=+1112.502887189 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "metrics-server-cert" not found Dec 09 11:49:46 crc kubenswrapper[4770]: E1209 11:49:46.262376 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:47.262357186 +0000 UTC m=+1112.503115705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "webhook-server-cert" not found Dec 09 11:49:46 crc kubenswrapper[4770]: I1209 11:49:46.464210 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" event={"ID":"cf58af47-dce9-43df-90cc-7c642e3b2f3b","Type":"ContainerStarted","Data":"26f2cedc4aa9e108fa71b030b94b64b54450b53d83b44f3a729d0934d4b97edc"} Dec 09 11:49:47 crc kubenswrapper[4770]: I1209 11:49:47.337435 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:47 crc kubenswrapper[4770]: I1209 11:49:47.338170 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.337661 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.338445 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:49.338425252 +0000 UTC m=+1114.579183781 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "webhook-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.338368 4770 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.338671 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:49.338638967 +0000 UTC m=+1114.579397536 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "metrics-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: I1209 11:49:47.544040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.546172 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.546271 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:51.546249996 +0000 UTC m=+1116.787008515 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: I1209 11:49:47.685512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.685801 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:47 crc kubenswrapper[4770]: E1209 11:49:47.685860 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert podName:86d62982-f66a-4184-8e15-9336bfca2b8f nodeName:}" failed. No retries permitted until 2025-12-09 11:49:51.685840215 +0000 UTC m=+1116.926598744 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" (UID: "86d62982-f66a-4184-8e15-9336bfca2b8f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.296728 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.343641 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.382974 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.404226 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.425033 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2"] Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.456691 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod211dd097_458c_435c_b3b0_b952dc866fac.slice/crio-2084706a4d085fad87a7f715dd85d20edf5270e05d96803baf95ebcbba633b41 WatchSource:0}: Error finding container 2084706a4d085fad87a7f715dd85d20edf5270e05d96803baf95ebcbba633b41: Status 404 returned error can't find the container with id 2084706a4d085fad87a7f715dd85d20edf5270e05d96803baf95ebcbba633b41 Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.477190 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.510994 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l"] Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.534483 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70ea2809_e1eb_4cf6_bd48_21ad18c031a4.slice/crio-e96ee75c4078eca40ab7236fbd1f038317a7a76b677242f7c377d8b83fec5bf8 WatchSource:0}: Error finding container e96ee75c4078eca40ab7236fbd1f038317a7a76b677242f7c377d8b83fec5bf8: Status 404 returned error can't find the container with id e96ee75c4078eca40ab7236fbd1f038317a7a76b677242f7c377d8b83fec5bf8 Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.535197 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.545321 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.556336 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd"] Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.569881 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxs8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vjw9d_openstack-operators(520a9036-0b49-4236-90cf-ada8df687ad9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.570397 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-xszh4"] Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.572289 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxs8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vjw9d_openstack-operators(520a9036-0b49-4236-90cf-ada8df687ad9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.574786 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" podUID="520a9036-0b49-4236-90cf-ada8df687ad9" Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.578273 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7af1b779_e5fd_40a2_8913_9a334a300efb.slice/crio-7570e84778ada3b4a9f9ce6df50fef724cb079ab157b0a3cf2a7890968bc1d19 WatchSource:0}: Error finding container 7570e84778ada3b4a9f9ce6df50fef724cb079ab157b0a3cf2a7890968bc1d19: Status 404 returned error can't find the container with id 7570e84778ada3b4a9f9ce6df50fef724cb079ab157b0a3cf2a7890968bc1d19 Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.584645 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" event={"ID":"2537825c-aa47-40ac-bb98-d9793449e9dd","Type":"ContainerStarted","Data":"98ef7e8c8dbd31dd229999c7377671974bb111971e1fc37d9bf802ebf4798c01"} Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.587580 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n49zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-bq5ts_openstack-operators(7af1b779-e5fd-40a2-8913-9a334a300efb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.587730 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts"] Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.590756 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n49zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-bq5ts_openstack-operators(7af1b779-e5fd-40a2-8913-9a334a300efb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.591837 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" podUID="7af1b779-e5fd-40a2-8913-9a334a300efb" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.592028 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" event={"ID":"70ea2809-e1eb-4cf6-bd48-21ad18c031a4","Type":"ContainerStarted","Data":"e96ee75c4078eca40ab7236fbd1f038317a7a76b677242f7c377d8b83fec5bf8"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.595488 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" event={"ID":"daa9d1b4-0867-4e4b-a5a4-037de4506c62","Type":"ContainerStarted","Data":"aea0b2bf5791584c1a8857abf58d4b91f3291cb8195070d2db089ae5e1f9f994"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.599686 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.602555 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" event={"ID":"3588fb02-b477-40dd-976e-f9b8e0508324","Type":"ContainerStarted","Data":"7ebef48c18cb1757dcaa9af518d8fbc60c9583c7d6167625ecfbe69280c4f517"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.606794 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-wljmn"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.609556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" event={"ID":"764304ee-eb30-434a-b884-f7e455576411","Type":"ContainerStarted","Data":"a0a713538c9f658fe9d9a270171477a5962e99234bcc2533fc51726cf4ffc7e2"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.629753 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" event={"ID":"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc","Type":"ContainerStarted","Data":"29218a149e5af93db13bf7162c116424e736853f699713294b7e761724ce4969"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.631988 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.643218 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" event={"ID":"7a873c67-c1eb-4681-b18e-25e1580bb11c","Type":"ContainerStarted","Data":"35da3a46764b88c13d599194d226ffa40a3996f3d401b2f5cccde1345a888b2c"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.646649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" event={"ID":"211dd097-458c-435c-b3b0-b952dc866fac","Type":"ContainerStarted","Data":"2084706a4d085fad87a7f715dd85d20edf5270e05d96803baf95ebcbba633b41"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.647290 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.648370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" event={"ID":"23265ff6-b252-4fc0-a3b4-6b59deea7c69","Type":"ContainerStarted","Data":"aa08dd1d82dd32a82842d3150be6e81a1cca28a32fa38d44101738bed101fc08"} Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.659475 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bggpm"] Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.660552 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd01836c_e495_43e0_8552_e17a41352a3d.slice/crio-92a1aed5ee7cd6c927d548a15c635979e828c0efdb150bbdde783ea3e74e053d WatchSource:0}: Error finding container 92a1aed5ee7cd6c927d548a15c635979e828c0efdb150bbdde783ea3e74e053d: Status 404 returned error can't find the container with id 92a1aed5ee7cd6c927d548a15c635979e828c0efdb150bbdde783ea3e74e053d Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.660739 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xktb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-wljmn_openstack-operators(7eea33f2-55bf-4542-82db-7cf987f29a63): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.660859 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" event={"ID":"303a0c89-f011-4953-ac5c-33929697ccd1","Type":"ContainerStarted","Data":"5de919c52d3bc4ef8b5b417f86ae5a840199254923e461bcc7d1066affd69a46"} Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.660871 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nl55g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-jv56l_openstack-operators(8c0e510c-adb0-43f4-b11d-05b378876a46): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.663509 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod049491b6_ec1f_421a_bfb3_bc7c63f27f8c.slice/crio-8f32e3e0c9f232188656ac46e383d4cb25166fc9f03a2e30583bd34e18564422 WatchSource:0}: Error finding container 8f32e3e0c9f232188656ac46e383d4cb25166fc9f03a2e30583bd34e18564422: Status 404 returned error can't find the container with id 8f32e3e0c9f232188656ac46e383d4cb25166fc9f03a2e30583bd34e18564422 Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.663570 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wgg7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-5jr9q_openstack-operators(fd01836c-e495-43e0-8552-e17a41352a3d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.663723 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xktb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-wljmn_openstack-operators(7eea33f2-55bf-4542-82db-7cf987f29a63): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.664049 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nl55g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-jv56l_openstack-operators(8c0e510c-adb0-43f4-b11d-05b378876a46): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.664863 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" podUID="7eea33f2-55bf-4542-82db-7cf987f29a63" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.665003 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" event={"ID":"c8cbcbfa-fe34-4fae-96b6-04a413418bf0","Type":"ContainerStarted","Data":"81df204a697a353ea99eb452c3dfdc4adaed10ae91788e09a621ad37a0734427"} Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.665162 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" podUID="8c0e510c-adb0-43f4-b11d-05b378876a46" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.665659 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wgg7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-5jr9q_openstack-operators(fd01836c-e495-43e0-8552-e17a41352a3d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.666748 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" podUID="fd01836c-e495-43e0-8552-e17a41352a3d" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.670313 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w8zjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-bggpm_openstack-operators(049491b6-ec1f-421a-bfb3-bc7c63f27f8c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.674408 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w8zjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-bggpm_openstack-operators(049491b6-ec1f-421a-bfb3-bc7c63f27f8c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.675884 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" podUID="049491b6-ec1f-421a-bfb3-bc7c63f27f8c" Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.769526 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h"] Dec 09 11:49:48 crc kubenswrapper[4770]: I1209 11:49:48.795017 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22"] Dec 09 11:49:48 crc kubenswrapper[4770]: W1209 11:49:48.822817 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3da14cb0_e4a9_4fe0_bf66_616a2ed438eb.slice/crio-54b3e1c1028934c44aef7491bd8ca1b57612f71775b12f4eec7ab501ad07f9bf WatchSource:0}: Error finding container 54b3e1c1028934c44aef7491bd8ca1b57612f71775b12f4eec7ab501ad07f9bf: Status 404 returned error can't find the container with id 54b3e1c1028934c44aef7491bd8ca1b57612f71775b12f4eec7ab501ad07f9bf Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.826778 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqjmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-b7x22_openstack-operators(3da14cb0-e4a9-4fe0-bf66-616a2ed438eb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.829494 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqjmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-b7x22_openstack-operators(3da14cb0-e4a9-4fe0-bf66-616a2ed438eb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 11:49:48 crc kubenswrapper[4770]: E1209 11:49:48.830745 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.373173 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.373467 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.373701 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.373735 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:53.373695355 +0000 UTC m=+1118.614454054 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "webhook-server-cert" not found Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.373880 4770 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.373982 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:53.373959002 +0000 UTC m=+1118.614717711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "metrics-server-cert" not found Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.690961 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" event={"ID":"7eea33f2-55bf-4542-82db-7cf987f29a63","Type":"ContainerStarted","Data":"d8175c395ea2ea7b162d69fbe3f5da77b79d58f0eed6948e208c704b55079257"} Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.701712 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" podUID="7eea33f2-55bf-4542-82db-7cf987f29a63" Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.703163 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" event={"ID":"8c0e510c-adb0-43f4-b11d-05b378876a46","Type":"ContainerStarted","Data":"b1c2a5a30511ab5546a480756caa6c2900d71d4c22f5cb895e3acfd524cbf7a4"} Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.706709 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" event={"ID":"520a9036-0b49-4236-90cf-ada8df687ad9","Type":"ContainerStarted","Data":"8297f6121ebbca8fb48c38c49af564041812e5a717a9091dec2e93c558b67145"} Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.710094 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" podUID="8c0e510c-adb0-43f4-b11d-05b378876a46" Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.710436 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" podUID="520a9036-0b49-4236-90cf-ada8df687ad9" Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.722216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" event={"ID":"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb","Type":"ContainerStarted","Data":"54b3e1c1028934c44aef7491bd8ca1b57612f71775b12f4eec7ab501ad07f9bf"} Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.742126 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.744737 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" event={"ID":"7af1b779-e5fd-40a2-8913-9a334a300efb","Type":"ContainerStarted","Data":"7570e84778ada3b4a9f9ce6df50fef724cb079ab157b0a3cf2a7890968bc1d19"} Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.751470 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" event={"ID":"30458aa5-ed86-491e-b2bd-353725caf57d","Type":"ContainerStarted","Data":"42d80621cb856c9075305526f89d0b01cec2e93b8c4796f1914c4e9801f33cf9"} Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.755727 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" podUID="7af1b779-e5fd-40a2-8913-9a334a300efb" Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.759727 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" event={"ID":"049491b6-ec1f-421a-bfb3-bc7c63f27f8c","Type":"ContainerStarted","Data":"8f32e3e0c9f232188656ac46e383d4cb25166fc9f03a2e30583bd34e18564422"} Dec 09 11:49:49 crc kubenswrapper[4770]: I1209 11:49:49.765974 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" event={"ID":"fd01836c-e495-43e0-8552-e17a41352a3d","Type":"ContainerStarted","Data":"92a1aed5ee7cd6c927d548a15c635979e828c0efdb150bbdde783ea3e74e053d"} Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.769933 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" podUID="049491b6-ec1f-421a-bfb3-bc7c63f27f8c" Dec 09 11:49:49 crc kubenswrapper[4770]: E1209 11:49:49.770599 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" podUID="fd01836c-e495-43e0-8552-e17a41352a3d" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.991741 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" podUID="8c0e510c-adb0-43f4-b11d-05b378876a46" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.993703 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" podUID="7af1b779-e5fd-40a2-8913-9a334a300efb" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.993850 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.994141 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" podUID="fd01836c-e495-43e0-8552-e17a41352a3d" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.994609 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" podUID="7eea33f2-55bf-4542-82db-7cf987f29a63" Dec 09 11:49:50 crc kubenswrapper[4770]: E1209 11:49:50.995816 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" podUID="520a9036-0b49-4236-90cf-ada8df687ad9" Dec 09 11:49:51 crc kubenswrapper[4770]: E1209 11:49:51.001503 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" podUID="049491b6-ec1f-421a-bfb3-bc7c63f27f8c" Dec 09 11:49:51 crc kubenswrapper[4770]: I1209 11:49:51.586622 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:51 crc kubenswrapper[4770]: E1209 11:49:51.586819 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:51 crc kubenswrapper[4770]: E1209 11:49:51.586881 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:49:59.58685935 +0000 UTC m=+1124.827617869 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:51 crc kubenswrapper[4770]: I1209 11:49:51.689035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:51 crc kubenswrapper[4770]: E1209 11:49:51.689458 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:51 crc kubenswrapper[4770]: E1209 11:49:51.689573 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert podName:86d62982-f66a-4184-8e15-9336bfca2b8f nodeName:}" failed. No retries permitted until 2025-12-09 11:49:59.689542081 +0000 UTC m=+1124.930300600 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" (UID: "86d62982-f66a-4184-8e15-9336bfca2b8f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 11:49:53 crc kubenswrapper[4770]: I1209 11:49:53.375853 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:53 crc kubenswrapper[4770]: I1209 11:49:53.376009 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:49:53 crc kubenswrapper[4770]: E1209 11:49:53.376199 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 11:49:53 crc kubenswrapper[4770]: E1209 11:49:53.376261 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:50:01.376243192 +0000 UTC m=+1126.617001711 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "webhook-server-cert" not found Dec 09 11:49:53 crc kubenswrapper[4770]: E1209 11:49:53.376553 4770 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 11:49:53 crc kubenswrapper[4770]: E1209 11:49:53.376704 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs podName:c57a5114-2563-4390-855b-402e73c907d6 nodeName:}" failed. No retries permitted until 2025-12-09 11:50:01.376671563 +0000 UTC m=+1126.617430282 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs") pod "openstack-operator-controller-manager-668858c49-vsvv8" (UID: "c57a5114-2563-4390-855b-402e73c907d6") : secret "metrics-server-cert" not found Dec 09 11:49:59 crc kubenswrapper[4770]: I1209 11:49:59.606361 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:49:59 crc kubenswrapper[4770]: E1209 11:49:59.607277 4770 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:59 crc kubenswrapper[4770]: E1209 11:49:59.607360 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert podName:d241ba8f-5573-466a-88f2-89ec9f14cc29 nodeName:}" failed. No retries permitted until 2025-12-09 11:50:15.607334874 +0000 UTC m=+1140.848093383 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert") pod "infra-operator-controller-manager-78d48bff9d-kf5v2" (UID: "d241ba8f-5573-466a-88f2-89ec9f14cc29") : secret "infra-operator-webhook-server-cert" not found Dec 09 11:49:59 crc kubenswrapper[4770]: I1209 11:49:59.708382 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:59 crc kubenswrapper[4770]: I1209 11:49:59.716016 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d62982-f66a-4184-8e15-9336bfca2b8f-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4\" (UID: \"86d62982-f66a-4184-8e15-9336bfca2b8f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:49:59 crc kubenswrapper[4770]: I1209 11:49:59.723517 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:50:01 crc kubenswrapper[4770]: I1209 11:50:01.462066 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:01 crc kubenswrapper[4770]: I1209 11:50:01.462727 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:01 crc kubenswrapper[4770]: I1209 11:50:01.467402 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:01 crc kubenswrapper[4770]: I1209 11:50:01.474471 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c57a5114-2563-4390-855b-402e73c907d6-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-vsvv8\" (UID: \"c57a5114-2563-4390-855b-402e73c907d6\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:01 crc kubenswrapper[4770]: I1209 11:50:01.654848 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:03 crc kubenswrapper[4770]: I1209 11:50:03.563801 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:50:05 crc kubenswrapper[4770]: E1209 11:50:05.074033 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 09 11:50:05 crc kubenswrapper[4770]: E1209 11:50:05.074277 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m74xq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-p69k8_openstack-operators(3588fb02-b477-40dd-976e-f9b8e0508324): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:06 crc kubenswrapper[4770]: E1209 11:50:06.141890 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 09 11:50:06 crc kubenswrapper[4770]: E1209 11:50:06.142510 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6ss9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-56gd2_openstack-operators(2537825c-aa47-40ac-bb98-d9793449e9dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:07 crc kubenswrapper[4770]: E1209 11:50:07.947575 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 09 11:50:07 crc kubenswrapper[4770]: E1209 11:50:07.947872 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-89l8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-h5w2l_openstack-operators(303a0c89-f011-4953-ac5c-33929697ccd1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:09 crc kubenswrapper[4770]: E1209 11:50:09.078091 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 09 11:50:09 crc kubenswrapper[4770]: E1209 11:50:09.078602 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-trxbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-xszh4_openstack-operators(70ea2809-e1eb-4cf6-bd48-21ad18c031a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:12 crc kubenswrapper[4770]: E1209 11:50:12.056084 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 09 11:50:12 crc kubenswrapper[4770]: E1209 11:50:12.056424 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w4g62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-w86cd_openstack-operators(764304ee-eb30-434a-b884-f7e455576411): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:13 crc kubenswrapper[4770]: E1209 11:50:13.266226 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 09 11:50:13 crc kubenswrapper[4770]: E1209 11:50:13.267047 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fwznj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-4lvc9_openstack-operators(23265ff6-b252-4fc0-a3b4-6b59deea7c69): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:15 crc kubenswrapper[4770]: I1209 11:50:15.687403 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:50:15 crc kubenswrapper[4770]: I1209 11:50:15.706846 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d241ba8f-5573-466a-88f2-89ec9f14cc29-cert\") pod \"infra-operator-controller-manager-78d48bff9d-kf5v2\" (UID: \"d241ba8f-5573-466a-88f2-89ec9f14cc29\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:50:15 crc kubenswrapper[4770]: I1209 11:50:15.877795 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r6nxf" Dec 09 11:50:15 crc kubenswrapper[4770]: I1209 11:50:15.884337 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:50:18 crc kubenswrapper[4770]: E1209 11:50:18.538976 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 09 11:50:18 crc kubenswrapper[4770]: E1209 11:50:18.539863 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jkc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-r89wl_openstack-operators(9a0fffd2-c599-4a15-b1f0-3c404d44a7bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:19 crc kubenswrapper[4770]: E1209 11:50:19.836666 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 09 11:50:19 crc kubenswrapper[4770]: E1209 11:50:19.837827 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqjmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-b7x22_openstack-operators(3da14cb0-e4a9-4fe0-bf66-616a2ed438eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:20 crc kubenswrapper[4770]: E1209 11:50:20.739000 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 09 11:50:20 crc kubenswrapper[4770]: E1209 11:50:20.739207 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xktb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-wljmn_openstack-operators(7eea33f2-55bf-4542-82db-7cf987f29a63): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:21 crc kubenswrapper[4770]: E1209 11:50:21.699989 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 09 11:50:21 crc kubenswrapper[4770]: E1209 11:50:21.700408 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wgg7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-5jr9q_openstack-operators(fd01836c-e495-43e0-8552-e17a41352a3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:22 crc kubenswrapper[4770]: E1209 11:50:22.637942 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 09 11:50:22 crc kubenswrapper[4770]: E1209 11:50:22.638190 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nl55g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-jv56l_openstack-operators(8c0e510c-adb0-43f4-b11d-05b378876a46): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:23 crc kubenswrapper[4770]: I1209 11:50:23.325613 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4"] Dec 09 11:50:24 crc kubenswrapper[4770]: E1209 11:50:24.915226 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8" Dec 09 11:50:24 crc kubenswrapper[4770]: E1209 11:50:24.915837 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n49zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-bq5ts_openstack-operators(7af1b779-e5fd-40a2-8913-9a334a300efb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:25 crc kubenswrapper[4770]: E1209 11:50:25.728466 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 09 11:50:25 crc kubenswrapper[4770]: E1209 11:50:25.728697 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v24pj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-h6w6h_openstack-operators(30458aa5-ed86-491e-b2bd-353725caf57d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:25 crc kubenswrapper[4770]: E1209 11:50:25.729876 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" podUID="30458aa5-ed86-491e-b2bd-353725caf57d" Dec 09 11:50:26 crc kubenswrapper[4770]: E1209 11:50:26.603283 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" podUID="30458aa5-ed86-491e-b2bd-353725caf57d" Dec 09 11:50:27 crc kubenswrapper[4770]: E1209 11:50:27.186508 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 09 11:50:27 crc kubenswrapper[4770]: E1209 11:50:27.187087 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxpq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-846ng_openstack-operators(7a873c67-c1eb-4681-b18e-25e1580bb11c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:27 crc kubenswrapper[4770]: I1209 11:50:27.613869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" event={"ID":"86d62982-f66a-4184-8e15-9336bfca2b8f","Type":"ContainerStarted","Data":"a08f6c588b07165d5e88d446a69f14e32fa1fec04e0d7bf7004bbc6470e362a6"} Dec 09 11:50:28 crc kubenswrapper[4770]: E1209 11:50:28.093553 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 09 11:50:28 crc kubenswrapper[4770]: E1209 11:50:28.093959 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxs8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vjw9d_openstack-operators(520a9036-0b49-4236-90cf-ada8df687ad9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:28 crc kubenswrapper[4770]: I1209 11:50:28.629821 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" event={"ID":"211dd097-458c-435c-b3b0-b952dc866fac","Type":"ContainerStarted","Data":"4bf1572867a3af0a3f26c0a4b157564450ec3ffaf46b2d0ee6e80a302cfbd408"} Dec 09 11:50:28 crc kubenswrapper[4770]: I1209 11:50:28.712036 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2"] Dec 09 11:50:28 crc kubenswrapper[4770]: I1209 11:50:28.822610 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8"] Dec 09 11:50:28 crc kubenswrapper[4770]: W1209 11:50:28.888725 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc57a5114_2563_4390_855b_402e73c907d6.slice/crio-684f421b34192fa0291f38d7ec56bf9a81a0e513509d8c47aca0de79e3701364 WatchSource:0}: Error finding container 684f421b34192fa0291f38d7ec56bf9a81a0e513509d8c47aca0de79e3701364: Status 404 returned error can't find the container with id 684f421b34192fa0291f38d7ec56bf9a81a0e513509d8c47aca0de79e3701364 Dec 09 11:50:29 crc kubenswrapper[4770]: I1209 11:50:29.797519 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" event={"ID":"d241ba8f-5573-466a-88f2-89ec9f14cc29","Type":"ContainerStarted","Data":"2a5a80c7337da4f56315bf122602b5ea0f3facb86416742de1032cfaf8ed7d69"} Dec 09 11:50:29 crc kubenswrapper[4770]: I1209 11:50:29.801103 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" event={"ID":"cf58af47-dce9-43df-90cc-7c642e3b2f3b","Type":"ContainerStarted","Data":"76bddf79ee041065145aad6aa10696f3f6486f1123cd675d72c8831fd805bddd"} Dec 09 11:50:29 crc kubenswrapper[4770]: I1209 11:50:29.803239 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" event={"ID":"c57a5114-2563-4390-855b-402e73c907d6","Type":"ContainerStarted","Data":"684f421b34192fa0291f38d7ec56bf9a81a0e513509d8c47aca0de79e3701364"} Dec 09 11:50:29 crc kubenswrapper[4770]: I1209 11:50:29.808813 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" event={"ID":"c8cbcbfa-fe34-4fae-96b6-04a413418bf0","Type":"ContainerStarted","Data":"acc44fdb5453d71db7178733d120eb79792239c5b22e94a541fc71398e392d8c"} Dec 09 11:50:30 crc kubenswrapper[4770]: I1209 11:50:30.831279 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" event={"ID":"daa9d1b4-0867-4e4b-a5a4-037de4506c62","Type":"ContainerStarted","Data":"7c75a43f07a704ce06c179fbe3dd6b62b997863a2e0d04dfd6ad9797fd105c67"} Dec 09 11:50:32 crc kubenswrapper[4770]: I1209 11:50:32.474039 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:50:32 crc kubenswrapper[4770]: I1209 11:50:32.474428 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:50:34 crc kubenswrapper[4770]: I1209 11:50:34.901477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" event={"ID":"c57a5114-2563-4390-855b-402e73c907d6","Type":"ContainerStarted","Data":"bcb928f737ef3ea514eae0a9f376c2a4a48ecaa6752f03403d36d642510f1830"} Dec 09 11:50:35 crc kubenswrapper[4770]: I1209 11:50:35.370353 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" podStartSLOduration=52.370318538 podStartE2EDuration="52.370318538s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:50:34.957165814 +0000 UTC m=+1160.197924423" watchObservedRunningTime="2025-12-09 11:50:35.370318538 +0000 UTC m=+1160.611077057" Dec 09 11:50:35 crc kubenswrapper[4770]: I1209 11:50:35.924241 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" event={"ID":"049491b6-ec1f-421a-bfb3-bc7c63f27f8c","Type":"ContainerStarted","Data":"900b616232eacb3b4b135b1c056b91490c29c3326cc489c882e4968a8b6efa19"} Dec 09 11:50:35 crc kubenswrapper[4770]: I1209 11:50:35.924483 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:41 crc kubenswrapper[4770]: I1209 11:50:41.661529 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-668858c49-vsvv8" Dec 09 11:50:47 crc kubenswrapper[4770]: E1209 11:50:47.822932 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:47 crc kubenswrapper[4770]: E1209 11:50:47.824259 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d4ggg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-5gj7z_openstack-operators(cf58af47-dce9-43df-90cc-7c642e3b2f3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:47 crc kubenswrapper[4770]: E1209 11:50:47.825552 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" podUID="cf58af47-dce9-43df-90cc-7c642e3b2f3b" Dec 09 11:50:48 crc kubenswrapper[4770]: I1209 11:50:48.190212 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:50:48 crc kubenswrapper[4770]: I1209 11:50:48.193986 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" Dec 09 11:50:49 crc kubenswrapper[4770]: E1209 11:50:49.800600 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48" Dec 09 11:50:49 crc kubenswrapper[4770]: E1209 11:50:49.801558 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:add611bf73d5aab1ac07ef665281ed0e5ad1aded495b8b32927aa2e726abb29a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:2f23894a78a13a0ae52fa2f8ae1e1b99282bebecd0cfda3db696e5d371097eaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:36946a77001110f391fb254ec77129803a6b7c34dacfa1a4c8c51aa8d23d57c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:dd58b29b5d88662a621c685c2b76fe8a71cc9e82aa85dff22a66182a6ceef3ae,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:fc47ed1c6249c9f6ef13ef1eac82d5a34819a715dea5117d33df0d0dc69ace8b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:e21d35c272d016f4dbd323dc827ee83538c96674adfb188e362aa652ce167b61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:c2ace235f775334be02d78928802b76309543e869cc6b4b55843ee546691e6c3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:be77cc58b87f299b42bb2cbe74f3f8d028b8c887851a53209441b60e1363aeb5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5a548c25fe3d02f7a042cb0a6d28fc8039a34c4a3b3d07aadda4aba3a926e777,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:41dc9cf27a902d9c7b392d730bd761cf3c391a548a841e9e4d38e1571f3c53bf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:174f8f712eb5fdda5061a1a68624befb27bbe766842653788583ec74c5ae506a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:df14f6de785b8aefc38ceb5b47088405224cfa914977c9ab811514cc77b08a67,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:b8d76f96b6f17a3318d089c0b5c0e6c292d969ab392cdcc708ec0f0188c953ae,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:43c55407c7c9b4141482533546e6570535373f7e36df374dfbbe388293c19dbf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:097816f289af117f14cd8ee1678a9635e8da6de4a1bde834d02199c4ef65c5c0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:3820c6110593bd1cef6a464a6fa8e01506644cfdf2233ea99c5201d47ebef645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:cfa46344aa9dc1d83053e4adacab02ec6313ffd222fa7d866002e02691cec7c6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:281668af8ed34c2464f3593d350cf7b695b41b81f40cc539ad74b7b65822afb9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:84319e5dd6569ea531e64b688557c2a2e20deb5225f3d349e402e34858f00fe7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:acb53e0e210562091843c212bc0cf5541daacd6f2bd18923430bae8c36578731,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:be6f4002842ebadf30d035721567a7e669f12a6eef8c00dc89030b3b08f3dd2c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:988635be61f6ed8c0d707622193b7efe8e9b1dc7effbf9b09d2db5ec593b59e7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:63e08752678a68571e1c54ceea42c113af493a04cdc22198a3713df7b53f87e5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:6741d06b0f1bbeb2968807dc5be45853cdd3dfb9cc7ea6ef23e909ae24f3cbf4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:1803a36d1a397a5595dddb4a2f791ab9443d3af97391a53928fa495ca7032d93,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:d163fcf801d67d9c67b2ae4368675b75714db7c531de842aad43979a888c5d57,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:15bf81d933a44128cb6f3264632a9563337eb3bfe82c4a33c746595467d3b0c3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:df38dbd6b3eccec2abaa8e3618a385405ccec1b73ae8c3573a138b0c961ed31f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:3a08e21338f651a90ee83ae46242b8c80c64488144f27a77848517049c3a8f5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:85729a662800e6b42ceb088545fed39a2ac58704b4a37fd540cdef3ebf9e59a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:ebeb4443ab9f9360925f7abd9c24b7a453390d678f79ed247d2042dcc6f9c3fc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:04bb4cd601b08034c6cba18e701fcd36026ec4340402ed710a0bbd09d8e4884d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:27b80783b7d4658d89dda9a09924e9ee472908a8fa1c86bcf3f773d17a4196e0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:8cb133c5a5551e1aa11ef3326149db1babbf00924d0ff493ebe3346b69fd4b5b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:13c3567176bb2d033f6c6b30e20404bd67a217e2537210bf222f3afe0c8619b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:60ac3446d57f1a97a6ca2d8e6584b00aa18704bc2707a7ac1a6a28c6d685d215,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:7e7788d1aae251e60f4012870140c65bce9760cd27feaeec5f65c42fe4ffce77,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:6a401117007514660c694248adce8136d83559caf1b38e475935335e09ac954a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:364d50f873551805782c23264570eff40e3807f35d9bccdd456515b4e31da488,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:2d72dd490576e0cb670d21a08420888f3758d64ed0cbd2ef8b9aa8488ad2ce40,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:96fdf7cddf31509ee63950a9d61320d0b01beb1212e28f37a6e872d6589ded22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:8b7534a2999075f919fc162d21f76026e8bf781913cc3d2ac07e484e9b2fc596,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:d65eaaea2ab02d63af9d8a106619908fa01a2e56bd6753edc5590e66e46270db,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:d042d7f91bafb002affff8cf750d694a0da129377255c502028528fe2280e790,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:a8faef9ea5e8ef8327b7fbb9b9cafc74c38c09c7e3b2365a7cad5eb49766f71d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:88aa46ea03a5584560806aa4b093584fda6b2f54c562005b72be2e3615688090,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:c08ecdfb7638c1897004347d835bdbabacff40a345f64c2b3111c377096bfa56,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:8b4025a4f30e83acc0b51ac063eea701006a302a1acbdec53f54b540270887f7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:4992f5ddbd20cca07e750846b2dbe7c51c5766c3002c388f8d8a158e347ec63d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:526afed30c44ef41d54d63a4f4db122bc603f775243ae350a59d2e0b5050076b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:22f097cb86b28ac48dc670ed7e0e841280bef1608f11b2b4536fbc2d2a6a90be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:20b3ad38accb9eb8849599280a263d3436a5af03d89645e5ec4508586297ffde,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:378ed518b68ea809cffa2ff7a93d51e52cfc53af14eedc978924fdabccef0325,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:8c3632033f8c004f31a1c7c57c5ca7b450a11e9170a220b8943b57f80717c70c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:3f746f7c6a8c48c0f4a800dcb4bc49bfbc4de4a9ca6a55d8f22bc515a92ea1d9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:e1f7bf105190c3cbbfcf0aeeb77a92d1466100ba8377221ed5eee228949e05bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:954b4c60705b229a968aba3b5b35ab02759378706103ed1189fae3e3316fac35,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:f2e0025727efb95efa65e6af6338ae3fc79bf61095d6d54931a0be8d7fe9acac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:2b4f8494513a3af102066fec5868ab167ac8664aceb2f0c639d7a0b60260a944,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:854a802357b4f565a366fce3bf29b20c1b768ec4ab7e822ef52dfc2fef000d2c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:194121c2d79401bd41f75428a437fe32a5806a6a160f7d80798ff66baed9afa5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:fa24ce4aa285e3632c86a53e8d0385d4c788d049da42dd06570ad9d44aae00de,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:df45459c449f64cc6471e98c0890ac00dcc77a940f85d4e7e9d9dd52990d65b3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:947c1bb9373b7d3f2acea104a5666e394c830111bf80d133f1fe7238e4d06f28,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:425ebddc9d6851ee9c730e67eaf43039943dc7937fb11332a41335a9114b2d44,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:bea03c7c34dc6ef8bc163e12a8940011b8feebc44a2efaaba2d3c4c6c515d6c8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:33f4e5f7a715d48482ec46a42267ea992fa268585303c4f1bd3cbea072a6348b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:a2280bc80b454dc9e5c95daf74b8a53d6f9e42fc16d45287e089fc41014fe1da,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:88d687a7bb593b2e61598b422baba84d67c114419590a6d83d15327d119ce208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:2635e02b99d380b2e547013c09c6c8da01bc89b3d3ce570e4d8f8656c7635b0e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:ac7fefe1c93839c7ccb2aaa0a18751df0e9f64a36a3b4cc1b81d82d7774b8b45,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a357cf166caaeea230f8a912aceb042e3170c5d680844e8f97b936baa10834ed,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:898a581a6b8e2b874d66f78a32fba831d891eb24af2cddb9ba93896484f3d09d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:f09b5a4f28f02ee1d8eece9045b9887551debc22cf9bedb9ff7623e1b007e187,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:9fb5fe24df2437e53d576aba5596a47659110eea87ae10da1386530bc501817b,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljt9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4_openstack-operators(86d62982-f66a-4184-8e15-9336bfca2b8f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.935338 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" podUID="cf58af47-dce9-43df-90cc-7c642e3b2f3b" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.950404 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.950588 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t9q5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-vvf9z_openstack-operators(211dd097-458c-435c-b3b0-b952dc866fac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.951831 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" podUID="211dd097-458c-435c-b3b0-b952dc866fac" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.953332 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.953546 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-trxbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-xszh4_openstack-operators(70ea2809-e1eb-4cf6-bd48-21ad18c031a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:52 crc kubenswrapper[4770]: E1209 11:50:52.957209 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" podUID="70ea2809-e1eb-4cf6-bd48-21ad18c031a4" Dec 09 11:50:53 crc kubenswrapper[4770]: I1209 11:50:53.225472 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:50:53 crc kubenswrapper[4770]: E1209 11:50:53.227391 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" podUID="cf58af47-dce9-43df-90cc-7c642e3b2f3b" Dec 09 11:50:53 crc kubenswrapper[4770]: E1209 11:50:53.228223 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" podUID="211dd097-458c-435c-b3b0-b952dc866fac" Dec 09 11:50:53 crc kubenswrapper[4770]: I1209 11:50:53.228740 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.008962 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.009162 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nl55g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-jv56l_openstack-operators(8c0e510c-adb0-43f4-b11d-05b378876a46): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.011184 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" podUID="8c0e510c-adb0-43f4-b11d-05b378876a46" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.011450 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.011706 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8cjct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-78d48bff9d-kf5v2_openstack-operators(d241ba8f-5573-466a-88f2-89ec9f14cc29): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.023635 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.023867 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fwznj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-4lvc9_openstack-operators(23265ff6-b252-4fc0-a3b4-6b59deea7c69): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.025028 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" podUID="23265ff6-b252-4fc0-a3b4-6b59deea7c69" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.027110 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.027534 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n49zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-bq5ts_openstack-operators(7af1b779-e5fd-40a2-8913-9a334a300efb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.028863 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" podUID="7af1b779-e5fd-40a2-8913-9a334a300efb" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.032950 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.033141 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxpq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-846ng_openstack-operators(7a873c67-c1eb-4681-b18e-25e1580bb11c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.034297 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" podUID="7a873c67-c1eb-4681-b18e-25e1580bb11c" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.034395 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.034534 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxs8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vjw9d_openstack-operators(520a9036-0b49-4236-90cf-ada8df687ad9): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.035328 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.035456 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xktb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-wljmn_openstack-operators(7eea33f2-55bf-4542-82db-7cf987f29a63): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.037083 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.038377 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cg6zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-zhc7c_openstack-operators(c8cbcbfa-fe34-4fae-96b6-04a413418bf0): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.037140 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" podUID="7eea33f2-55bf-4542-82db-7cf987f29a63" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.037208 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.038550 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zrwzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-96cv7_openstack-operators(daa9d1b4-0867-4e4b-a5a4-037de4506c62): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.037124 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" podUID="520a9036-0b49-4236-90cf-ada8df687ad9" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.039625 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" podUID="daa9d1b4-0867-4e4b-a5a4-037de4506c62" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.039712 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" podUID="c8cbcbfa-fe34-4fae-96b6-04a413418bf0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.119862 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.120076 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wgg7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-5jr9q_openstack-operators(fd01836c-e495-43e0-8552-e17a41352a3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.121778 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" podUID="fd01836c-e495-43e0-8552-e17a41352a3d" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.147930 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.148126 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jkc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-r89wl_openstack-operators(9a0fffd2-c599-4a15-b1f0-3c404d44a7bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.149462 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" podUID="9a0fffd2-c599-4a15-b1f0-3c404d44a7bc" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.174448 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.174808 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqjmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-b7x22_openstack-operators(3da14cb0-e4a9-4fe0-bf66-616a2ed438eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.176062 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" Dec 09 11:50:54 crc kubenswrapper[4770]: I1209 11:50:54.231312 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:50:54 crc kubenswrapper[4770]: I1209 11:50:54.237481 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.596625 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" podUID="d241ba8f-5573-466a-88f2-89ec9f14cc29" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.615465 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" podUID="2537825c-aa47-40ac-bb98-d9793449e9dd" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.636211 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" podUID="3588fb02-b477-40dd-976e-f9b8e0508324" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.648459 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" podUID="303a0c89-f011-4953-ac5c-33929697ccd1" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.722336 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" podUID="86d62982-f66a-4184-8e15-9336bfca2b8f" Dec 09 11:50:54 crc kubenswrapper[4770]: E1209 11:50:54.787915 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" podUID="764304ee-eb30-434a-b884-f7e455576411" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.258748 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" event={"ID":"23265ff6-b252-4fc0-a3b4-6b59deea7c69","Type":"ContainerStarted","Data":"7802b7714a7133c441e5ef87958019d37363bd40a54c4f314a0d6375315b7057"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.261186 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" event={"ID":"3588fb02-b477-40dd-976e-f9b8e0508324","Type":"ContainerStarted","Data":"9b7279d60a8d21942431777ed82bcb30cd58c08fe284fab4bfcccb21ed55ba76"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.308767 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" event={"ID":"70ea2809-e1eb-4cf6-bd48-21ad18c031a4","Type":"ContainerStarted","Data":"d19fb89781b3d18c0ea18f8387ca539b94d57efdfd3d5604a7174677645c1ef1"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.308831 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" event={"ID":"70ea2809-e1eb-4cf6-bd48-21ad18c031a4","Type":"ContainerStarted","Data":"63da0a6704bc04c521c61b9d6ab2f04f87d72f46dadc1e4f5fe342b50a93dbf3"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.312877 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.325969 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" event={"ID":"c8cbcbfa-fe34-4fae-96b6-04a413418bf0","Type":"ContainerStarted","Data":"dca947274f5d48966828edcb83f00a572bd202e4cab116c213d95e5f2c29a63e"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.326658 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.345407 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.367638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" event={"ID":"daa9d1b4-0867-4e4b-a5a4-037de4506c62","Type":"ContainerStarted","Data":"7d000bd457adcd898430791c737dc4dbe9cb3b4f95e340ffb8fac68acaf9265e"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.367688 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" event={"ID":"764304ee-eb30-434a-b884-f7e455576411","Type":"ContainerStarted","Data":"edd12b675c1a2ea6f3ecb645a84aaf5bacc25a73e8439429ecb8bacfb8261c45"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.379081 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" event={"ID":"7a873c67-c1eb-4681-b18e-25e1580bb11c","Type":"ContainerStarted","Data":"a3b33ec6be7bbee4f1bb1d0a0bd0c5c615d417fefd474655b459728787445f08"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.387008 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" event={"ID":"30458aa5-ed86-491e-b2bd-353725caf57d","Type":"ContainerStarted","Data":"358caa5934add22b2ef474fe92f559c338b31d876bf8a3edfb46af47a8724a30"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.392717 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" podStartSLOduration=6.751783426 podStartE2EDuration="1m12.392693357s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.550305382 +0000 UTC m=+1113.791063901" lastFinishedPulling="2025-12-09 11:50:54.191215313 +0000 UTC m=+1179.431973832" observedRunningTime="2025-12-09 11:50:55.372632577 +0000 UTC m=+1180.613391106" watchObservedRunningTime="2025-12-09 11:50:55.392693357 +0000 UTC m=+1180.633451876" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.402395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" event={"ID":"2537825c-aa47-40ac-bb98-d9793449e9dd","Type":"ContainerStarted","Data":"117a93883b3e149281cf9c4688ba50959f126e197d2bc75d2cacabeddb47fb10"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.424814 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" event={"ID":"86d62982-f66a-4184-8e15-9336bfca2b8f","Type":"ContainerStarted","Data":"5fa758f3a168f379da5668692a5d13d2a6298c73108716ec7bf42fe43d210b14"} Dec 09 11:50:55 crc kubenswrapper[4770]: E1209 11:50:55.428756 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" podUID="86d62982-f66a-4184-8e15-9336bfca2b8f" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.442610 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" event={"ID":"303a0c89-f011-4953-ac5c-33929697ccd1","Type":"ContainerStarted","Data":"7c1f15ec90f788f9bba9dfdd3ccceb0521de1af807d2db607712da9ad310c925"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.486304 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" event={"ID":"d241ba8f-5573-466a-88f2-89ec9f14cc29","Type":"ContainerStarted","Data":"49d716454ceea3daf1932306a0defe6fd8f7bb48dfd5dab387f236f1ec678bc2"} Dec 09 11:50:55 crc kubenswrapper[4770]: E1209 11:50:55.502719 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab\\\"\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" podUID="d241ba8f-5573-466a-88f2-89ec9f14cc29" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.518233 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-zhc7c" podStartSLOduration=40.28149056 podStartE2EDuration="1m13.518198558s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.450688209 +0000 UTC m=+1113.691446728" lastFinishedPulling="2025-12-09 11:50:21.687396207 +0000 UTC m=+1146.928154726" observedRunningTime="2025-12-09 11:50:55.417649361 +0000 UTC m=+1180.658407890" watchObservedRunningTime="2025-12-09 11:50:55.518198558 +0000 UTC m=+1180.758957087" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.529247 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" event={"ID":"211dd097-458c-435c-b3b0-b952dc866fac","Type":"ContainerStarted","Data":"9728f536728d93564d451aa9e58ef17bce856ae80a24cb6f378c8e92bcc8d17b"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.534011 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" event={"ID":"049491b6-ec1f-421a-bfb3-bc7c63f27f8c","Type":"ContainerStarted","Data":"05cabfe558ffc68ca3b51ddff64b5cd0adc3e0ba1c300cf43607a6caa9fe2c7b"} Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.534597 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.540842 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.572141 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-96cv7" podStartSLOduration=39.274718228 podStartE2EDuration="1m12.572114648s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.389764281 +0000 UTC m=+1113.630522800" lastFinishedPulling="2025-12-09 11:50:21.687160701 +0000 UTC m=+1146.927919220" observedRunningTime="2025-12-09 11:50:55.485815455 +0000 UTC m=+1180.726573974" watchObservedRunningTime="2025-12-09 11:50:55.572114648 +0000 UTC m=+1180.812873187" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.646041 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bggpm" podStartSLOduration=7.126918824 podStartE2EDuration="1m12.646019677s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.670096778 +0000 UTC m=+1113.910855297" lastFinishedPulling="2025-12-09 11:50:54.189197631 +0000 UTC m=+1179.429956150" observedRunningTime="2025-12-09 11:50:55.609830077 +0000 UTC m=+1180.850588606" watchObservedRunningTime="2025-12-09 11:50:55.646019677 +0000 UTC m=+1180.886778206" Dec 09 11:50:55 crc kubenswrapper[4770]: I1209 11:50:55.856799 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h6w6h" podStartSLOduration=6.456077643 podStartE2EDuration="1m11.856778586s" podCreationTimestamp="2025-12-09 11:49:44 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.790784716 +0000 UTC m=+1114.031543235" lastFinishedPulling="2025-12-09 11:50:54.191485649 +0000 UTC m=+1179.432244178" observedRunningTime="2025-12-09 11:50:55.854859307 +0000 UTC m=+1181.095617826" watchObservedRunningTime="2025-12-09 11:50:55.856778586 +0000 UTC m=+1181.097537105" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.022981 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vvf9z" podStartSLOduration=40.828357813 podStartE2EDuration="1m14.02295408s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.492598045 +0000 UTC m=+1113.733356574" lastFinishedPulling="2025-12-09 11:50:21.687194322 +0000 UTC m=+1146.927952841" observedRunningTime="2025-12-09 11:50:56.019446261 +0000 UTC m=+1181.260204790" watchObservedRunningTime="2025-12-09 11:50:56.02295408 +0000 UTC m=+1181.263712599" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.569287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" event={"ID":"764304ee-eb30-434a-b884-f7e455576411","Type":"ContainerStarted","Data":"0cac7e094159ccfd610cfffb004d19c9ac0636423dd9c4494d83e79f22a7d043"} Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.570548 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.584396 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" event={"ID":"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc","Type":"ContainerStarted","Data":"7d888fabfc91fb6ad0654fc5a453ff5c925bdee2781b084a98905f6318131ddf"} Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.584469 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" event={"ID":"9a0fffd2-c599-4a15-b1f0-3c404d44a7bc","Type":"ContainerStarted","Data":"6979ee65c265a4dcb6b72804938b2c2eb4e4a879cf0eba24fcf3dcb3d16af952"} Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.585547 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.598083 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" event={"ID":"7a873c67-c1eb-4681-b18e-25e1580bb11c","Type":"ContainerStarted","Data":"79a1d64d73a1d9d149a7c4da2a7e3df7f713a8d89b512827f50a9f7ac0f3458c"} Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.599081 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.603125 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" event={"ID":"23265ff6-b252-4fc0-a3b4-6b59deea7c69","Type":"ContainerStarted","Data":"779122e13e115f994ea9e73e4cef7e7777cfd4c646a016a4f0b6d3a2668d487a"} Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.603196 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.624663 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" event={"ID":"3588fb02-b477-40dd-976e-f9b8e0508324","Type":"ContainerStarted","Data":"69db689cbf2ccb5dc974ca222d94d53dce897be237c543e0e1e68d4a0299e2c3"} Dec 09 11:50:56 crc kubenswrapper[4770]: E1209 11:50:56.629543 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" podUID="86d62982-f66a-4184-8e15-9336bfca2b8f" Dec 09 11:50:56 crc kubenswrapper[4770]: E1209 11:50:56.629771 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab\\\"\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" podUID="d241ba8f-5573-466a-88f2-89ec9f14cc29" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.748502 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" podStartSLOduration=6.391675743 podStartE2EDuration="1m13.748480866s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.532997482 +0000 UTC m=+1113.773755991" lastFinishedPulling="2025-12-09 11:50:55.889802605 +0000 UTC m=+1181.130561114" observedRunningTime="2025-12-09 11:50:56.74276919 +0000 UTC m=+1181.983527709" watchObservedRunningTime="2025-12-09 11:50:56.748480866 +0000 UTC m=+1181.989239395" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.801145 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" podStartSLOduration=8.360367559 podStartE2EDuration="1m14.801124834s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.395083656 +0000 UTC m=+1113.635842175" lastFinishedPulling="2025-12-09 11:50:54.835840931 +0000 UTC m=+1180.076599450" observedRunningTime="2025-12-09 11:50:56.792342601 +0000 UTC m=+1182.033101120" watchObservedRunningTime="2025-12-09 11:50:56.801124834 +0000 UTC m=+1182.041883353" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.834485 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" podStartSLOduration=8.476754198 podStartE2EDuration="1m14.834462682s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.529183225 +0000 UTC m=+1113.769941734" lastFinishedPulling="2025-12-09 11:50:54.886891699 +0000 UTC m=+1180.127650218" observedRunningTime="2025-12-09 11:50:56.830974633 +0000 UTC m=+1182.071733152" watchObservedRunningTime="2025-12-09 11:50:56.834462682 +0000 UTC m=+1182.075221201" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.909337 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" podStartSLOduration=7.369252292 podStartE2EDuration="1m14.909309694s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.348200674 +0000 UTC m=+1113.588959193" lastFinishedPulling="2025-12-09 11:50:55.888258076 +0000 UTC m=+1181.129016595" observedRunningTime="2025-12-09 11:50:56.906441851 +0000 UTC m=+1182.147200370" watchObservedRunningTime="2025-12-09 11:50:56.909309694 +0000 UTC m=+1182.150068213" Dec 09 11:50:56 crc kubenswrapper[4770]: I1209 11:50:56.963398 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" podStartSLOduration=8.602115645 podStartE2EDuration="1m14.963375809s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.528934839 +0000 UTC m=+1113.769693358" lastFinishedPulling="2025-12-09 11:50:54.890195003 +0000 UTC m=+1180.130953522" observedRunningTime="2025-12-09 11:50:56.96222669 +0000 UTC m=+1182.202985219" watchObservedRunningTime="2025-12-09 11:50:56.963375809 +0000 UTC m=+1182.204134328" Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.634281 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" event={"ID":"303a0c89-f011-4953-ac5c-33929697ccd1","Type":"ContainerStarted","Data":"281baeb1c8c0e6cac46e93de55ce2c92b19fb4d3fadcd6be659e503a13afa7eb"} Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.635512 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.635881 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" event={"ID":"2537825c-aa47-40ac-bb98-d9793449e9dd","Type":"ContainerStarted","Data":"bcc5503ba14b006f89bd1c50ce0ab5e8442319ffa91d264caa8871db551b58fd"} Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.636331 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.670997 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" podStartSLOduration=8.065512883 podStartE2EDuration="1m15.670976688s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.478283401 +0000 UTC m=+1113.719041920" lastFinishedPulling="2025-12-09 11:50:56.083747206 +0000 UTC m=+1181.324505725" observedRunningTime="2025-12-09 11:50:57.662956695 +0000 UTC m=+1182.903715214" watchObservedRunningTime="2025-12-09 11:50:57.670976688 +0000 UTC m=+1182.911735207" Dec 09 11:50:57 crc kubenswrapper[4770]: I1209 11:50:57.689857 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" podStartSLOduration=8.04139019 podStartE2EDuration="1m15.689841438s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.395846535 +0000 UTC m=+1113.636605054" lastFinishedPulling="2025-12-09 11:50:56.044297783 +0000 UTC m=+1181.285056302" observedRunningTime="2025-12-09 11:50:57.687455797 +0000 UTC m=+1182.928214336" watchObservedRunningTime="2025-12-09 11:50:57.689841438 +0000 UTC m=+1182.930599957" Dec 09 11:50:58 crc kubenswrapper[4770]: I1209 11:50:58.811687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:51:02 crc kubenswrapper[4770]: I1209 11:51:02.474616 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:51:02 crc kubenswrapper[4770]: I1209 11:51:02.475236 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.021317 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-h5w2l" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.124563 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-r89wl" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.149030 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-p69k8" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.178509 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-56gd2" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.240439 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-4lvc9" Dec 09 11:51:03 crc kubenswrapper[4770]: I1209 11:51:03.512522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-846ng" Dec 09 11:51:04 crc kubenswrapper[4770]: I1209 11:51:04.422531 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-xszh4" Dec 09 11:51:04 crc kubenswrapper[4770]: I1209 11:51:04.475308 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-w86cd" Dec 09 11:51:07 crc kubenswrapper[4770]: I1209 11:51:07.174151 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" event={"ID":"8c0e510c-adb0-43f4-b11d-05b378876a46","Type":"ContainerStarted","Data":"0dde52e79c501fd42454a5bb844291faacaa4ee8f2fcbb9393baa0aa190e8126"} Dec 09 11:51:07 crc kubenswrapper[4770]: I1209 11:51:07.184210 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" event={"ID":"cf58af47-dce9-43df-90cc-7c642e3b2f3b","Type":"ContainerStarted","Data":"2e0530818bbcb1bf2c2994258a60e142318a752e2d0cdfbeccf0bf0e4b5b6ab6"} Dec 09 11:51:07 crc kubenswrapper[4770]: I1209 11:51:07.192333 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" event={"ID":"7af1b779-e5fd-40a2-8913-9a334a300efb","Type":"ContainerStarted","Data":"d26e45c5e7972afebb8b4af58d08e309e64e3ff106ddfe871b252e69a0cab056"} Dec 09 11:51:07 crc kubenswrapper[4770]: I1209 11:51:07.206587 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5gj7z" podStartSLOduration=50.232158487 podStartE2EDuration="1m25.206571131s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:45.758015354 +0000 UTC m=+1110.998773873" lastFinishedPulling="2025-12-09 11:50:20.732427998 +0000 UTC m=+1145.973186517" observedRunningTime="2025-12-09 11:51:07.205482593 +0000 UTC m=+1192.446241102" watchObservedRunningTime="2025-12-09 11:51:07.206571131 +0000 UTC m=+1192.447329650" Dec 09 11:51:08 crc kubenswrapper[4770]: I1209 11:51:08.252176 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" event={"ID":"8c0e510c-adb0-43f4-b11d-05b378876a46","Type":"ContainerStarted","Data":"0669dcceed6fcd239bdcbbc8c465994617b635080e71522fe117230fde899908"} Dec 09 11:51:08 crc kubenswrapper[4770]: I1209 11:51:08.252557 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:51:08 crc kubenswrapper[4770]: I1209 11:51:08.273738 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" podStartSLOduration=8.244986995 podStartE2EDuration="1m26.27371436s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.660674208 +0000 UTC m=+1113.901432727" lastFinishedPulling="2025-12-09 11:51:06.689401573 +0000 UTC m=+1191.930160092" observedRunningTime="2025-12-09 11:51:08.268107557 +0000 UTC m=+1193.508866096" watchObservedRunningTime="2025-12-09 11:51:08.27371436 +0000 UTC m=+1193.514472879" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.672599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" event={"ID":"86d62982-f66a-4184-8e15-9336bfca2b8f","Type":"ContainerStarted","Data":"0ce9c30d21c1b489ac886ba6e083647e38a3c6d3fe8cf63331a6caee8e5e064a"} Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.674423 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.685080 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" event={"ID":"7af1b779-e5fd-40a2-8913-9a334a300efb","Type":"ContainerStarted","Data":"de3b935638a7bdcf85a474f85b1c2044288d6d2bbc99361571b86a65d649dc6b"} Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.685975 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.687994 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" event={"ID":"fd01836c-e495-43e0-8552-e17a41352a3d","Type":"ContainerStarted","Data":"b2935dbd0ceaa9cc92811d7c26361b0303380becf80c2a767a767f467819bf14"} Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.690360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" event={"ID":"7eea33f2-55bf-4542-82db-7cf987f29a63","Type":"ContainerStarted","Data":"267ce8116b5269cf2322d08c6d5372e02c3bdee887c006dd69302b02188fbeeb"} Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.690413 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" event={"ID":"7eea33f2-55bf-4542-82db-7cf987f29a63","Type":"ContainerStarted","Data":"eb1bb209d9bc29cb3767bb9f82d277d548535bdf50e49b2082dc0c79b84c6e2d"} Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.690717 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.735167 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" podStartSLOduration=45.209625682 podStartE2EDuration="1m27.735144148s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:50:27.338590927 +0000 UTC m=+1152.579349446" lastFinishedPulling="2025-12-09 11:51:09.864109393 +0000 UTC m=+1195.104867912" observedRunningTime="2025-12-09 11:51:10.732587753 +0000 UTC m=+1195.973346272" watchObservedRunningTime="2025-12-09 11:51:10.735144148 +0000 UTC m=+1195.975902667" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.759260 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" podStartSLOduration=9.657761456 podStartE2EDuration="1m27.759236171s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.587321223 +0000 UTC m=+1113.828079742" lastFinishedPulling="2025-12-09 11:51:06.688795938 +0000 UTC m=+1191.929554457" observedRunningTime="2025-12-09 11:51:10.754664304 +0000 UTC m=+1195.995422823" watchObservedRunningTime="2025-12-09 11:51:10.759236171 +0000 UTC m=+1195.999994690" Dec 09 11:51:10 crc kubenswrapper[4770]: I1209 11:51:10.774339 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" podStartSLOduration=8.875069577 podStartE2EDuration="1m27.774320574s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.660575885 +0000 UTC m=+1113.901334404" lastFinishedPulling="2025-12-09 11:51:07.559826882 +0000 UTC m=+1192.800585401" observedRunningTime="2025-12-09 11:51:10.771122003 +0000 UTC m=+1196.011880522" watchObservedRunningTime="2025-12-09 11:51:10.774320574 +0000 UTC m=+1196.015079103" Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.709579 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" event={"ID":"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb","Type":"ContainerStarted","Data":"0d7933ff61a52f6d8e5e329f7d9c41918dd69cf6bf3de387875a1497805f5daa"} Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.710209 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" event={"ID":"3da14cb0-e4a9-4fe0-bf66-616a2ed438eb","Type":"ContainerStarted","Data":"5fed56b4bfeee27135fefd2db7c0a78f75b801e6716326535e62c212ec416b3b"} Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.710714 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.712631 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" event={"ID":"fd01836c-e495-43e0-8552-e17a41352a3d","Type":"ContainerStarted","Data":"c20209f5c9a5478d6c7155c616f8f95c585ebc06fe429acf9fb71b659745fb7e"} Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.713249 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.714888 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" event={"ID":"520a9036-0b49-4236-90cf-ada8df687ad9","Type":"ContainerStarted","Data":"fbe165b2e2a7b4728d6e0310f65bb45682b75845c42fa7b9791928e76693691f"} Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.714959 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" event={"ID":"520a9036-0b49-4236-90cf-ada8df687ad9","Type":"ContainerStarted","Data":"c286890ba6431c72955b8e042878dca149a74ea7ba1b6979b753de18d5d3389d"} Dec 09 11:51:11 crc kubenswrapper[4770]: I1209 11:51:11.715789 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:51:12 crc kubenswrapper[4770]: I1209 11:51:12.155438 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podStartSLOduration=7.629319458 podStartE2EDuration="1m29.155410276s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.826545595 +0000 UTC m=+1114.067304124" lastFinishedPulling="2025-12-09 11:51:10.352636423 +0000 UTC m=+1195.593394942" observedRunningTime="2025-12-09 11:51:12.149373983 +0000 UTC m=+1197.390132522" watchObservedRunningTime="2025-12-09 11:51:12.155410276 +0000 UTC m=+1197.396168795" Dec 09 11:51:12 crc kubenswrapper[4770]: I1209 11:51:12.184160 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" podStartSLOduration=8.837329306 podStartE2EDuration="1m30.184120166s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.663400407 +0000 UTC m=+1113.904158926" lastFinishedPulling="2025-12-09 11:51:10.010191267 +0000 UTC m=+1195.250949786" observedRunningTime="2025-12-09 11:51:12.181380376 +0000 UTC m=+1197.422138895" watchObservedRunningTime="2025-12-09 11:51:12.184120166 +0000 UTC m=+1197.424878685" Dec 09 11:51:12 crc kubenswrapper[4770]: I1209 11:51:12.215331 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" podStartSLOduration=7.774267753 podStartE2EDuration="1m29.215313909s" podCreationTimestamp="2025-12-09 11:49:43 +0000 UTC" firstStartedPulling="2025-12-09 11:49:48.569731236 +0000 UTC m=+1113.810489755" lastFinishedPulling="2025-12-09 11:51:10.010777392 +0000 UTC m=+1195.251535911" observedRunningTime="2025-12-09 11:51:12.2070816 +0000 UTC m=+1197.447840129" watchObservedRunningTime="2025-12-09 11:51:12.215313909 +0000 UTC m=+1197.456072428" Dec 09 11:51:12 crc kubenswrapper[4770]: I1209 11:51:12.734426 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-bq5ts" Dec 09 11:51:13 crc kubenswrapper[4770]: I1209 11:51:13.740592 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" event={"ID":"d241ba8f-5573-466a-88f2-89ec9f14cc29","Type":"ContainerStarted","Data":"a909c56d5510d524726419e7183f2b2a23cacf46bb3df772fb97a73a21b8e6f7"} Dec 09 11:51:13 crc kubenswrapper[4770]: I1209 11:51:13.742099 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:51:13 crc kubenswrapper[4770]: I1209 11:51:13.998805 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" podStartSLOduration=48.718901105 podStartE2EDuration="1m31.99878276s" podCreationTimestamp="2025-12-09 11:49:42 +0000 UTC" firstStartedPulling="2025-12-09 11:50:28.897194702 +0000 UTC m=+1154.137953221" lastFinishedPulling="2025-12-09 11:51:12.177076357 +0000 UTC m=+1197.417834876" observedRunningTime="2025-12-09 11:51:13.994940292 +0000 UTC m=+1199.235698831" watchObservedRunningTime="2025-12-09 11:51:13.99878276 +0000 UTC m=+1199.239541269" Dec 09 11:51:14 crc kubenswrapper[4770]: I1209 11:51:14.189039 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-wljmn" Dec 09 11:51:15 crc kubenswrapper[4770]: I1209 11:51:15.039645 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jv56l" Dec 09 11:51:19 crc kubenswrapper[4770]: I1209 11:51:19.731220 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4" Dec 09 11:51:24 crc kubenswrapper[4770]: I1209 11:51:24.064384 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-5jr9q" Dec 09 11:51:24 crc kubenswrapper[4770]: I1209 11:51:24.074289 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vjw9d" Dec 09 11:51:24 crc kubenswrapper[4770]: I1209 11:51:24.451425 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" Dec 09 11:51:25 crc kubenswrapper[4770]: I1209 11:51:25.891474 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-kf5v2" Dec 09 11:51:32 crc kubenswrapper[4770]: I1209 11:51:32.473991 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:51:32 crc kubenswrapper[4770]: I1209 11:51:32.474333 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:51:32 crc kubenswrapper[4770]: I1209 11:51:32.474382 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:51:32 crc kubenswrapper[4770]: I1209 11:51:32.475049 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:51:32 crc kubenswrapper[4770]: I1209 11:51:32.475100 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c" gracePeriod=600 Dec 09 11:51:33 crc kubenswrapper[4770]: I1209 11:51:33.041851 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c" exitCode=0 Dec 09 11:51:33 crc kubenswrapper[4770]: I1209 11:51:33.041929 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c"} Dec 09 11:51:33 crc kubenswrapper[4770]: I1209 11:51:33.042218 4770 scope.go:117] "RemoveContainer" containerID="6cfcd66816cbbc56831ecf2144f5b9d68700b9ef841df4897036f813b08a8d24" Dec 09 11:51:34 crc kubenswrapper[4770]: I1209 11:51:34.053972 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f"} Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.562601 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.565394 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.567700 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.567753 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkzqt\" (UniqueName: \"kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.569943 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.570434 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.570713 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.570749 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7wq8z" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.573687 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.622583 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.629202 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.633592 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.636426 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.668793 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkzqt\" (UniqueName: \"kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.669812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.671243 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.690736 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkzqt\" (UniqueName: \"kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt\") pod \"dnsmasq-dns-84bb9d8bd9-9fdcx\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.770793 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.770996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzx66\" (UniqueName: \"kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.771068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.872815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzx66\" (UniqueName: \"kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.872880 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.872958 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.874011 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.874138 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.892748 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzx66\" (UniqueName: \"kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66\") pod \"dnsmasq-dns-5f854695bc-psdt9\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.894456 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:51:46 crc kubenswrapper[4770]: I1209 11:51:46.944462 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:51:48 crc kubenswrapper[4770]: I1209 11:51:48.879853 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:51:48 crc kubenswrapper[4770]: W1209 11:51:48.891185 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6448998e_1823_4155_8ed4_6e3a343856f6.slice/crio-9402e01f06dc1cd36e39b099761ef81b986701ab0ee03a448793bb1b824d92c3 WatchSource:0}: Error finding container 9402e01f06dc1cd36e39b099761ef81b986701ab0ee03a448793bb1b824d92c3: Status 404 returned error can't find the container with id 9402e01f06dc1cd36e39b099761ef81b986701ab0ee03a448793bb1b824d92c3 Dec 09 11:51:48 crc kubenswrapper[4770]: I1209 11:51:48.943262 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:51:48 crc kubenswrapper[4770]: W1209 11:51:48.950985 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5726c3e7_d186_4cd0_89e9_abc20a4115e7.slice/crio-b17047f4a3e3f2b5fd362f4dcc7d1f300934cc4d5a0f6547cee87bfd9935e146 WatchSource:0}: Error finding container b17047f4a3e3f2b5fd362f4dcc7d1f300934cc4d5a0f6547cee87bfd9935e146: Status 404 returned error can't find the container with id b17047f4a3e3f2b5fd362f4dcc7d1f300934cc4d5a0f6547cee87bfd9935e146 Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.263105 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.318066 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.319675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.334205 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.515304 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.515427 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.515602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq5tf\" (UniqueName: \"kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.863744 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq5tf\" (UniqueName: \"kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.863940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.863976 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.866148 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.869128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.894139 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" event={"ID":"5726c3e7-d186-4cd0-89e9-abc20a4115e7","Type":"ContainerStarted","Data":"b17047f4a3e3f2b5fd362f4dcc7d1f300934cc4d5a0f6547cee87bfd9935e146"} Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.900518 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" event={"ID":"6448998e-1823-4155-8ed4-6e3a343856f6","Type":"ContainerStarted","Data":"9402e01f06dc1cd36e39b099761ef81b986701ab0ee03a448793bb1b824d92c3"} Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.909626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq5tf\" (UniqueName: \"kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf\") pod \"dnsmasq-dns-744ffd65bc-m6j8n\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.947794 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.967397 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.996758 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:51:49 crc kubenswrapper[4770]: I1209 11:51:49.998792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.021449 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.067356 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.067423 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.067471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlns\" (UniqueName: \"kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.746166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.746278 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlns\" (UniqueName: \"kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.746498 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.747924 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:50 crc kubenswrapper[4770]: I1209 11:51:50.749776 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.158740 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlns\" (UniqueName: \"kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns\") pod \"dnsmasq-dns-95f5f6995-svm7p\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.484512 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.852278 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.855032 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.855174 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.868805 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.869311 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.872993 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.873270 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.873433 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.873597 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.873987 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hvrph" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.932769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.932839 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.932862 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.932884 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.932987 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933078 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxvnd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933099 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933150 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933183 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933222 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:51 crc kubenswrapper[4770]: I1209 11:51:51.933303 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.022017 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.024119 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.028705 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rmbjp" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.030645 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.032775 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.034624 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.034847 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035216 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035328 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035432 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035568 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035756 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxvnd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.035967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.036124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.038922 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.116742 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.117034 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.117348 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.134064 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.136482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.139007 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.139672 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.140403 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.141527 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.155289 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.162097 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.167233 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.170847 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.175039 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.180019 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xxtr4" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.180461 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.180085 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.181739 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.182739 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.227370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.227992 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxvnd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.228289 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.232806 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.232889 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.239842 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.239942 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.240033 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.240072 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.246319 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.246386 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.246411 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.246468 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.247111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.247170 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpnmk\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.247240 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396472 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396518 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396555 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396615 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396696 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcddb\" (UniqueName: \"kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396734 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396791 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396817 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396844 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396870 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.396984 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpnmk\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.397016 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.397040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.397081 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.398225 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.398461 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.402452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.405665 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.406004 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.406375 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.411586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.412283 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.417546 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.418306 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.459743 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.479834 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpnmk\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498530 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcddb\" (UniqueName: \"kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498646 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498678 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498704 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.498880 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.499043 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.502154 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.507211 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.510293 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.511428 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.511475 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.508535 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.513342 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.515405 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.542433 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcddb\" (UniqueName: \"kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.546717 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.551253 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:51:52 crc kubenswrapper[4770]: W1209 11:51:52.958783 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e7a3769_1def_45f0_80b8_f73f0c8868bd.slice/crio-80a00bbdaf5cb5c2dffa4fa18dd3f7db0f584c61a3a72c7255f9671b99722800 WatchSource:0}: Error finding container 80a00bbdaf5cb5c2dffa4fa18dd3f7db0f584c61a3a72c7255f9671b99722800: Status 404 returned error can't find the container with id 80a00bbdaf5cb5c2dffa4fa18dd3f7db0f584c61a3a72c7255f9671b99722800 Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.973505 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " pod="openstack/openstack-galera-0" Dec 09 11:51:52 crc kubenswrapper[4770]: I1209 11:51:52.999542 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:51:53 crc kubenswrapper[4770]: I1209 11:51:53.168865 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.059033 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.078259 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.079973 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.084830 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.108839 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.108877 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rbr5j" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.109266 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.120782 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.149385 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" event={"ID":"7150d0a7-b185-4942-bd70-3232d267f23a","Type":"ContainerStarted","Data":"d70c11b1730ed8683d1f9ed7f886cb596cc954d6ccac99be0f83ff9e3ef5fda9"} Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.175346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" event={"ID":"7e7a3769-1def-45f0-80b8-f73f0c8868bd","Type":"ContainerStarted","Data":"80a00bbdaf5cb5c2dffa4fa18dd3f7db0f584c61a3a72c7255f9671b99722800"} Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.176812 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281495 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281554 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbf9p\" (UniqueName: \"kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281632 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281647 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281669 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281701 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.281740 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.365295 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.373858 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388052 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388707 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388788 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbf9p\" (UniqueName: \"kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388860 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.388975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.389003 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.389035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.389085 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.391850 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.392118 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.392789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.393253 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nwhdp" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.393440 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.395183 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.396013 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.396131 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.420490 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbf9p\" (UniqueName: \"kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.425012 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.441139 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.442639 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.497882 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.497952 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn62r\" (UniqueName: \"kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.498199 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.498277 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.498326 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.509741 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.605132 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.605260 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.605319 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.605376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.605403 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn62r\" (UniqueName: \"kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.607027 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.610141 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.612591 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:54 crc kubenswrapper[4770]: I1209 11:51:54.921352 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 11:51:55 crc kubenswrapper[4770]: I1209 11:51:54.971323 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn62r\" (UniqueName: \"kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:55 crc kubenswrapper[4770]: I1209 11:51:55.130161 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " pod="openstack/memcached-0" Dec 09 11:51:56 crc kubenswrapper[4770]: I1209 11:51:56.035193 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerStarted","Data":"372b4b00e2cc2abd1ce34b82c880ec351754b64a39d8320cda277799057ff294"} Dec 09 11:51:56 crc kubenswrapper[4770]: I1209 11:51:56.035457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerStarted","Data":"269faed7cb2f222c44ee3f00b36632affafba0f9d4af49831d837ec797f8e210"} Dec 09 11:51:56 crc kubenswrapper[4770]: I1209 11:51:56.035475 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerStarted","Data":"98e3471866b7b3ea7ac43432961287c5a59b6b0f6249f5401b19933d817b43a7"} Dec 09 11:51:56 crc kubenswrapper[4770]: I1209 11:51:56.148962 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 11:51:56 crc kubenswrapper[4770]: I1209 11:51:56.598849 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.395138 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.526722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerStarted","Data":"8351da7f332ac8f54987f847f02c25d2625de68f0947f11791e8e48f707cdbd6"} Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.562572 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f352a11c-1887-4736-98cb-4eccf3086e97","Type":"ContainerStarted","Data":"7a853b3079ae327a5e2751c66ecf028c4a7e55fd1f1c6894af589afedfaf48e0"} Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.660075 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.661777 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.670608 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.767610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qqwc\" (UniqueName: \"kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc\") pod \"kube-state-metrics-0\" (UID: \"76483913-5fec-49d0-8d19-916a024ec7fd\") " pod="openstack/kube-state-metrics-0" Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.768352 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4ww8c" Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.868786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qqwc\" (UniqueName: \"kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc\") pod \"kube-state-metrics-0\" (UID: \"76483913-5fec-49d0-8d19-916a024ec7fd\") " pod="openstack/kube-state-metrics-0" Dec 09 11:51:57 crc kubenswrapper[4770]: I1209 11:51:57.936257 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qqwc\" (UniqueName: \"kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc\") pod \"kube-state-metrics-0\" (UID: \"76483913-5fec-49d0-8d19-916a024ec7fd\") " pod="openstack/kube-state-metrics-0" Dec 09 11:51:58 crc kubenswrapper[4770]: I1209 11:51:58.130480 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.555707 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.563120 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.572076 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-x6cbq" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.572451 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.572661 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584219 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584408 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584444 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj5lm\" (UniqueName: \"kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584554 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.584703 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.641059 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692016 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj5lm\" (UniqueName: \"kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692803 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692180 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692964 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.692986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.693022 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.693101 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.693706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.705980 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.714530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.747879 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj5lm\" (UniqueName: \"kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.752297 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle\") pod \"ovn-controller-xhnq4\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " pod="openstack/ovn-controller-xhnq4" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.776751 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.778854 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804168 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804200 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804253 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804386 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804409 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804438 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4575p\" (UniqueName: \"kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.804599 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.908974 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.911078 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.912879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.913274 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4575p\" (UniqueName: \"kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.913472 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.913577 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.913774 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.913187 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.914526 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.914928 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.924761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.938487 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4575p\" (UniqueName: \"kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p\") pod \"ovn-controller-ovs-gct95\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.953364 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:51:59 crc kubenswrapper[4770]: I1209 11:51:59.980792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.217492 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.429030 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.431158 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.438465 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.442746 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.452616 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.453017 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-r9gjh" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.453274 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.453399 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655316 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655399 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655430 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntppd\" (UniqueName: \"kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655451 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655698 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655796 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.655934 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.757316 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntppd\" (UniqueName: \"kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758321 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758388 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758465 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758505 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758555 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.758591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.762313 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.762601 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.762891 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.768075 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.769299 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.790246 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.799372 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.809693 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntppd\" (UniqueName: \"kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.875586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:00 crc kubenswrapper[4770]: I1209 11:52:00.980714 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76483913-5fec-49d0-8d19-916a024ec7fd","Type":"ContainerStarted","Data":"0010a467df1fe6e371c53ff6a681be6a0e157914c226d73cb140cda8dd5d73cc"} Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.066268 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.764122 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.969486 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.972304 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.982574 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.983884 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.984080 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.986306 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-mzqrv" Dec 09 11:52:01 crc kubenswrapper[4770]: I1209 11:52:01.992146 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.099119 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105549 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105598 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqfnq\" (UniqueName: \"kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105617 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105681 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105740 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105763 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.105802 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.208685 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.207967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.208948 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210148 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210469 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqfnq\" (UniqueName: \"kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210513 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210598 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210669 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.210693 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.211013 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.212211 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.212318 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.218376 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.223021 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.235403 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.236125 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqfnq\" (UniqueName: \"kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.236869 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.258222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: I1209 11:52:02.355460 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:02 crc kubenswrapper[4770]: W1209 11:52:02.603498 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod518c58bd_4a2f_439f_b5f5_0eda21434884.slice/crio-80f870d5a6fc083f41c08a37e38b9c189c2fffdeffd28ec9149331a846b31a2a WatchSource:0}: Error finding container 80f870d5a6fc083f41c08a37e38b9c189c2fffdeffd28ec9149331a846b31a2a: Status 404 returned error can't find the container with id 80f870d5a6fc083f41c08a37e38b9c189c2fffdeffd28ec9149331a846b31a2a Dec 09 11:52:03 crc kubenswrapper[4770]: I1209 11:52:03.084395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerStarted","Data":"80f870d5a6fc083f41c08a37e38b9c189c2fffdeffd28ec9149331a846b31a2a"} Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.370863 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.373297 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.388826 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.407039 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542148 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542482 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542505 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542543 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.542593 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrrh\" (UniqueName: \"kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645073 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrrh\" (UniqueName: \"kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645222 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645261 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645289 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645317 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645347 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.645795 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.647510 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.647606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.653557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.669553 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.752699 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrrh\" (UniqueName: \"kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh\") pod \"ovn-controller-metrics-jfg8z\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.829055 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.847722 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.858419 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.863781 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.865149 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.866788 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.870194 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.870281 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.870310 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.870722 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkj96\" (UniqueName: \"kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.973846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.973985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.974021 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.974136 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkj96\" (UniqueName: \"kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.976607 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.977183 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:04 crc kubenswrapper[4770]: I1209 11:52:04.977370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:05 crc kubenswrapper[4770]: I1209 11:52:05.015045 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkj96\" (UniqueName: \"kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96\") pod \"dnsmasq-dns-7878659675-g2r4x\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:05 crc kubenswrapper[4770]: I1209 11:52:05.614936 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:08 crc kubenswrapper[4770]: I1209 11:52:08.010523 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-n9pcb" podUID="844a755c-0f78-40a6-93f3-a388dce59835" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.54:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:52:08 crc kubenswrapper[4770]: I1209 11:52:08.032355 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-n9pcb" podUID="844a755c-0f78-40a6-93f3-a388dce59835" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.54:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:52:08 crc kubenswrapper[4770]: I1209 11:52:08.064437 4770 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.81194958s: [/var/lib/containers/storage/overlay/c53adf464a0be179c177e87452a029e0ffc7e08e063707b29af29d21267460d7/diff /var/log/pods/openshift-console_downloads-7954f5f757-xwjv4_e69f8571-c321-4ff0-9fd1-fb67b1176230/download-server/2.log]; will not log again for this container unless duration exceeds 2s Dec 09 11:52:08 crc kubenswrapper[4770]: W1209 11:52:08.708610 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5969e8b4_92d6_4ef0_8a28_3a3b8ea10be2.slice/crio-4cbc671a83e48d5acdf23fe813dbec0fda6b957ad84cbf364cb5f7e1cda1889f WatchSource:0}: Error finding container 4cbc671a83e48d5acdf23fe813dbec0fda6b957ad84cbf364cb5f7e1cda1889f: Status 404 returned error can't find the container with id 4cbc671a83e48d5acdf23fe813dbec0fda6b957ad84cbf364cb5f7e1cda1889f Dec 09 11:52:09 crc kubenswrapper[4770]: I1209 11:52:09.099281 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4" event={"ID":"f754e609-fd35-4c45-bfe8-71c659a22cdb","Type":"ContainerStarted","Data":"a1f58f7451278f7de20c41adf37fd92ef7d6e8e3761163d4d91c67e6e0ff8b1d"} Dec 09 11:52:09 crc kubenswrapper[4770]: I1209 11:52:09.100405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerStarted","Data":"4cbc671a83e48d5acdf23fe813dbec0fda6b957ad84cbf364cb5f7e1cda1889f"} Dec 09 11:52:16 crc kubenswrapper[4770]: I1209 11:52:16.616247 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:52:17 crc kubenswrapper[4770]: I1209 11:52:17.179779 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerStarted","Data":"fc5566166462f989c3e85ae3adf50733b43e9d8583eb013a176ba72ae6dfbff5"} Dec 09 11:52:26 crc kubenswrapper[4770]: I1209 11:52:26.933802 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:52:26 crc kubenswrapper[4770]: I1209 11:52:26.944561 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:52:28 crc kubenswrapper[4770]: W1209 11:52:28.476511 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81bf1b9e_0c6e_43d8_81df_a16fb7aa59a7.slice/crio-98393f4f7c1f9414a8e905e6561d3fdffa792dee7064ef74bf062533fcdb7091 WatchSource:0}: Error finding container 98393f4f7c1f9414a8e905e6561d3fdffa792dee7064ef74bf062533fcdb7091: Status 404 returned error can't find the container with id 98393f4f7c1f9414a8e905e6561d3fdffa792dee7064ef74bf062533fcdb7091 Dec 09 11:52:29 crc kubenswrapper[4770]: I1209 11:52:29.327959 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-g2r4x" event={"ID":"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7","Type":"ContainerStarted","Data":"98393f4f7c1f9414a8e905e6561d3fdffa792dee7064ef74bf062533fcdb7091"} Dec 09 11:52:30 crc kubenswrapper[4770]: W1209 11:52:30.751223 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcabab723_8add_4dd0_b0cc_551f30a039d3.slice/crio-a5082f92fc1418c0900e41864024dfdfaaf0407045931ba63005855d9fe50f5f WatchSource:0}: Error finding container a5082f92fc1418c0900e41864024dfdfaaf0407045931ba63005855d9fe50f5f: Status 404 returned error can't find the container with id a5082f92fc1418c0900e41864024dfdfaaf0407045931ba63005855d9fe50f5f Dec 09 11:52:31 crc kubenswrapper[4770]: I1209 11:52:31.354229 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jfg8z" event={"ID":"cabab723-8add-4dd0-b0cc-551f30a039d3","Type":"ContainerStarted","Data":"a5082f92fc1418c0900e41864024dfdfaaf0407045931ba63005855d9fe50f5f"} Dec 09 11:52:39 crc kubenswrapper[4770]: E1209 11:52:39.915438 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13" Dec 09 11:52:39 crc kubenswrapper[4770]: E1209 11:52:39.916019 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hcddb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(d6e15893-6d15-4fa7-abc0-33697cc9b4cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:39 crc kubenswrapper[4770]: E1209 11:52:39.917236 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" Dec 09 11:52:40 crc kubenswrapper[4770]: E1209 11:52:40.431460 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13\\\"\"" pod="openstack/openstack-galera-0" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" Dec 09 11:52:40 crc kubenswrapper[4770]: E1209 11:52:40.987624 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d" Dec 09 11:52:40 crc kubenswrapper[4770]: E1209 11:52:40.988318 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxvnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(03fef9de-3f78-48b3-9079-9dc87184f803): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:40 crc kubenswrapper[4770]: E1209 11:52:40.989488 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.074504 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.074878 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jpnmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(b596f863-b8d8-4fb4-93e3-45bdce78f1b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.076120 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.441968 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.442077 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:e733252aab7f4bc0efbdd712bcd88e44c5498bf1773dba843bc9dcfac324fe3d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.710324 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.710580 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n56dh59chddh58h64bh9fhcdh565hd4h9bh667h5f8h7dh66h59chc7h5d7hd6h659hd4h7h59dhb8h99h54bh78h597hdfh89h664h5c7h5b8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mn62r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(f352a11c-1887-4736-98cb-4eccf3086e97): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:41 crc kubenswrapper[4770]: E1209 11:52:41.711843 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.001276 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.001354 4770 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.001523 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6qqwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(76483913-5fec-49d0-8d19-916a024ec7fd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.002715 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.575119 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:e47191ba776414b781b3e27b856ab45a03b9480c7dc2b1addb939608794882dc\\\"\"" pod="openstack/memcached-0" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" Dec 09 11:52:42 crc kubenswrapper[4770]: E1209 11:52:42.576021 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" Dec 09 11:52:43 crc kubenswrapper[4770]: E1209 11:52:43.433870 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13" Dec 09 11:52:43 crc kubenswrapper[4770]: E1209 11:52:43.434434 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cbf9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(f17f6cc7-50ef-45f1-9552-edc09f1e6ecb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:43 crc kubenswrapper[4770]: E1209 11:52:43.435623 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" Dec 09 11:52:43 crc kubenswrapper[4770]: E1209 11:52:43.585141 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:ed0f8ba03f3ce47a32006d730c3049455325eb2c3b98b9fd6b3fb9901004df13\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.132778 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.133063 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:df45459c449f64cc6471e98c0890ac00dcc77a940f85d4e7e9d9dd52990d65b3" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.133117 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzx66,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5f854695bc-psdt9_openstack(6448998e-1823-4155-8ed4-6e3a343856f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.133266 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:df45459c449f64cc6471e98c0890ac00dcc77a940f85d4e7e9d9dd52990d65b3,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncfh67bh9fh98h5cfh5b4h67ch99h5cdhc5h74h64ch96h5dhb8h565h54fh574h5d5h568h98h65h566hd5h55bhbch59bh9fh65dh67chd5h5fcq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4575p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-gct95_openstack(5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.134393 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" podUID="6448998e-1823-4155-8ed4-6e3a343856f6" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.134441 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.136956 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.137108 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nq5tf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-744ffd65bc-m6j8n_openstack(7e7a3769-1def-45f0-80b8-f73f0c8868bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.138391 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" podUID="7e7a3769-1def-45f0-80b8-f73f0c8868bd" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.147795 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.148074 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ltlns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-95f5f6995-svm7p_openstack(7150d0a7-b185-4942-bd70-3232d267f23a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.149232 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" podUID="7150d0a7-b185-4942-bd70-3232d267f23a" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.186338 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.186568 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dkzqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84bb9d8bd9-9fdcx_openstack(5726c3e7-d186-4cd0-89e9-abc20a4115e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.188078 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" podUID="5726c3e7-d186-4cd0-89e9-abc20a4115e7" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.359217 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:fa24ce4aa285e3632c86a53e8d0385d4c788d049da42dd06570ad9d44aae00de" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.359423 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:fa24ce4aa285e3632c86a53e8d0385d4c788d049da42dd06570ad9d44aae00de,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncfh67bh9fh98h5cfh5b4h67ch99h5cdhc5h74h64ch96h5dhb8h565h54fh574h5d5h568h98h65h566hd5h55bhbch59bh9fh65dh67chd5h5fcq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bj5lm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-xhnq4_openstack(f754e609-fd35-4c45-bfe8-71c659a22cdb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.360708 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.593015 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:df45459c449f64cc6471e98c0890ac00dcc77a940f85d4e7e9d9dd52990d65b3\\\"\"" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.594768 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:fa24ce4aa285e3632c86a53e8d0385d4c788d049da42dd06570ad9d44aae00de\\\"\"" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" Dec 09 11:52:44 crc kubenswrapper[4770]: E1209 11:52:44.594851 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33\\\"\"" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" podUID="7150d0a7-b185-4942-bd70-3232d267f23a" Dec 09 11:52:46 crc kubenswrapper[4770]: E1209 11:52:46.082519 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:947c1bb9373b7d3f2acea104a5666e394c830111bf80d133f1fe7238e4d06f28" Dec 09 11:52:46 crc kubenswrapper[4770]: E1209 11:52:46.083229 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:947c1bb9373b7d3f2acea104a5666e394c830111bf80d133f1fe7238e4d06f28,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n596hf6h66fh595hbh568h549h7chf7hcch57bh54ch596hfch59fh56bh5c7h575h597hbh54bh644h5f9h66bh596h668h58chf6h545h74h595h548q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ntppd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(518c58bd-4a2f-439f-b5f5-0eda21434884): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.283494 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.290944 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.306158 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350457 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzx66\" (UniqueName: \"kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66\") pod \"6448998e-1823-4155-8ed4-6e3a343856f6\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350532 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config\") pod \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350587 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config\") pod \"6448998e-1823-4155-8ed4-6e3a343856f6\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350645 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config\") pod \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350661 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq5tf\" (UniqueName: \"kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf\") pod \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350725 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc\") pod \"6448998e-1823-4155-8ed4-6e3a343856f6\" (UID: \"6448998e-1823-4155-8ed4-6e3a343856f6\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350788 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkzqt\" (UniqueName: \"kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt\") pod \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\" (UID: \"5726c3e7-d186-4cd0-89e9-abc20a4115e7\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.350842 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc\") pod \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\" (UID: \"7e7a3769-1def-45f0-80b8-f73f0c8868bd\") " Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.351309 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config" (OuterVolumeSpecName: "config") pod "6448998e-1823-4155-8ed4-6e3a343856f6" (UID: "6448998e-1823-4155-8ed4-6e3a343856f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.351365 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config" (OuterVolumeSpecName: "config") pod "7e7a3769-1def-45f0-80b8-f73f0c8868bd" (UID: "7e7a3769-1def-45f0-80b8-f73f0c8868bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.351587 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config" (OuterVolumeSpecName: "config") pod "5726c3e7-d186-4cd0-89e9-abc20a4115e7" (UID: "5726c3e7-d186-4cd0-89e9-abc20a4115e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.351662 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e7a3769-1def-45f0-80b8-f73f0c8868bd" (UID: "7e7a3769-1def-45f0-80b8-f73f0c8868bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.352139 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6448998e-1823-4155-8ed4-6e3a343856f6" (UID: "6448998e-1823-4155-8ed4-6e3a343856f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.356638 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf" (OuterVolumeSpecName: "kube-api-access-nq5tf") pod "7e7a3769-1def-45f0-80b8-f73f0c8868bd" (UID: "7e7a3769-1def-45f0-80b8-f73f0c8868bd"). InnerVolumeSpecName "kube-api-access-nq5tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.360221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66" (OuterVolumeSpecName: "kube-api-access-vzx66") pod "6448998e-1823-4155-8ed4-6e3a343856f6" (UID: "6448998e-1823-4155-8ed4-6e3a343856f6"). InnerVolumeSpecName "kube-api-access-vzx66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.369102 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt" (OuterVolumeSpecName: "kube-api-access-dkzqt") pod "5726c3e7-d186-4cd0-89e9-abc20a4115e7" (UID: "5726c3e7-d186-4cd0-89e9-abc20a4115e7"). InnerVolumeSpecName "kube-api-access-dkzqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453670 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5726c3e7-d186-4cd0-89e9-abc20a4115e7-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453716 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq5tf\" (UniqueName: \"kubernetes.io/projected/7e7a3769-1def-45f0-80b8-f73f0c8868bd-kube-api-access-nq5tf\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453733 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453748 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkzqt\" (UniqueName: \"kubernetes.io/projected/5726c3e7-d186-4cd0-89e9-abc20a4115e7-kube-api-access-dkzqt\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453760 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453774 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzx66\" (UniqueName: \"kubernetes.io/projected/6448998e-1823-4155-8ed4-6e3a343856f6-kube-api-access-vzx66\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453792 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7a3769-1def-45f0-80b8-f73f0c8868bd-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.453804 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6448998e-1823-4155-8ed4-6e3a343856f6-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.610656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" event={"ID":"5726c3e7-d186-4cd0-89e9-abc20a4115e7","Type":"ContainerDied","Data":"b17047f4a3e3f2b5fd362f4dcc7d1f300934cc4d5a0f6547cee87bfd9935e146"} Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.610779 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-9fdcx" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.617000 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.617052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-m6j8n" event={"ID":"7e7a3769-1def-45f0-80b8-f73f0c8868bd","Type":"ContainerDied","Data":"80a00bbdaf5cb5c2dffa4fa18dd3f7db0f584c61a3a72c7255f9671b99722800"} Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.622770 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" event={"ID":"6448998e-1823-4155-8ed4-6e3a343856f6","Type":"ContainerDied","Data":"9402e01f06dc1cd36e39b099761ef81b986701ab0ee03a448793bb1b824d92c3"} Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.622948 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-psdt9" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.692615 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.707470 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-m6j8n"] Dec 09 11:52:46 crc kubenswrapper[4770]: E1209 11:52:46.752771 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.761368 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.795409 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-9fdcx"] Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.832278 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:52:46 crc kubenswrapper[4770]: I1209 11:52:46.840071 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-psdt9"] Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.349145 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5726c3e7-d186-4cd0-89e9-abc20a4115e7" path="/var/lib/kubelet/pods/5726c3e7-d186-4cd0-89e9-abc20a4115e7/volumes" Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.349961 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6448998e-1823-4155-8ed4-6e3a343856f6" path="/var/lib/kubelet/pods/6448998e-1823-4155-8ed4-6e3a343856f6/volumes" Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.350374 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7a3769-1def-45f0-80b8-f73f0c8868bd" path="/var/lib/kubelet/pods/7e7a3769-1def-45f0-80b8-f73f0c8868bd/volumes" Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.637440 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerStarted","Data":"7b634173148f824de10487babb4e2f965f1000bfa7ef0c55f90a98000dbccbff"} Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.637506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerStarted","Data":"f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67"} Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.640476 4770 generic.go:334] "Generic (PLEG): container finished" podID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerID="63b6b28586679902472eef833c7b0c5e21e38bbedc017c3ddc0df2decab51165" exitCode=0 Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.641166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-g2r4x" event={"ID":"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7","Type":"ContainerDied","Data":"63b6b28586679902472eef833c7b0c5e21e38bbedc017c3ddc0df2decab51165"} Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.647766 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jfg8z" event={"ID":"cabab723-8add-4dd0-b0cc-551f30a039d3","Type":"ContainerStarted","Data":"33128ce32a57707b0adfe4148bebb6dd61f023516c99ba4c5d553f50864934f7"} Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.656751 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerStarted","Data":"5525f70a6a808175e62c919c6f6e56ec54293e10c519fefa5b6d1b5a45deea6e"} Dec 09 11:52:47 crc kubenswrapper[4770]: E1209 11:52:47.658529 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:947c1bb9373b7d3f2acea104a5666e394c830111bf80d133f1fe7238e4d06f28\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.668823 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.171946743 podStartE2EDuration="47.66880178s" podCreationTimestamp="2025-12-09 11:52:00 +0000 UTC" firstStartedPulling="2025-12-09 11:52:16.997193618 +0000 UTC m=+1262.237952137" lastFinishedPulling="2025-12-09 11:52:46.494048655 +0000 UTC m=+1291.734807174" observedRunningTime="2025-12-09 11:52:47.66566508 +0000 UTC m=+1292.906423599" watchObservedRunningTime="2025-12-09 11:52:47.66880178 +0000 UTC m=+1292.909560299" Dec 09 11:52:47 crc kubenswrapper[4770]: I1209 11:52:47.712690 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-jfg8z" podStartSLOduration=27.954286693 podStartE2EDuration="43.712665282s" podCreationTimestamp="2025-12-09 11:52:04 +0000 UTC" firstStartedPulling="2025-12-09 11:52:30.754092203 +0000 UTC m=+1275.994850722" lastFinishedPulling="2025-12-09 11:52:46.512470782 +0000 UTC m=+1291.753229311" observedRunningTime="2025-12-09 11:52:47.705233813 +0000 UTC m=+1292.945992332" watchObservedRunningTime="2025-12-09 11:52:47.712665282 +0000 UTC m=+1292.953423801" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.110594 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.154675 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.156477 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.158986 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.197794 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.290818 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.290894 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.291057 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.291107 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.291177 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph6f7\" (UniqueName: \"kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.393260 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.393365 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.393433 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.393545 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph6f7\" (UniqueName: \"kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.393682 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.394602 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.394836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.395815 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.396111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.429216 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph6f7\" (UniqueName: \"kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7\") pod \"dnsmasq-dns-586b989cdc-6mk5d\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.502118 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.531831 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.597028 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc\") pod \"7150d0a7-b185-4942-bd70-3232d267f23a\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.597191 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltlns\" (UniqueName: \"kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns\") pod \"7150d0a7-b185-4942-bd70-3232d267f23a\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.597297 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config\") pod \"7150d0a7-b185-4942-bd70-3232d267f23a\" (UID: \"7150d0a7-b185-4942-bd70-3232d267f23a\") " Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.597795 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config" (OuterVolumeSpecName: "config") pod "7150d0a7-b185-4942-bd70-3232d267f23a" (UID: "7150d0a7-b185-4942-bd70-3232d267f23a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.597974 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7150d0a7-b185-4942-bd70-3232d267f23a" (UID: "7150d0a7-b185-4942-bd70-3232d267f23a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.600465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns" (OuterVolumeSpecName: "kube-api-access-ltlns") pod "7150d0a7-b185-4942-bd70-3232d267f23a" (UID: "7150d0a7-b185-4942-bd70-3232d267f23a"). InnerVolumeSpecName "kube-api-access-ltlns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.675363 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-g2r4x" event={"ID":"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7","Type":"ContainerStarted","Data":"67c7f2fa7c544e10472b827f869f6a0dbb8df243afedca85c04542d0ccbb42ad"} Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.675484 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.676724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" event={"ID":"7150d0a7-b185-4942-bd70-3232d267f23a","Type":"ContainerDied","Data":"d70c11b1730ed8683d1f9ed7f886cb596cc954d6ccac99be0f83ff9e3ef5fda9"} Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.676773 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-svm7p" Dec 09 11:52:48 crc kubenswrapper[4770]: E1209 11:52:48.682770 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:947c1bb9373b7d3f2acea104a5666e394c830111bf80d133f1fe7238e4d06f28\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.697846 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7878659675-g2r4x" podStartSLOduration=26.725439176 podStartE2EDuration="44.6978269s" podCreationTimestamp="2025-12-09 11:52:04 +0000 UTC" firstStartedPulling="2025-12-09 11:52:28.497944229 +0000 UTC m=+1273.738702748" lastFinishedPulling="2025-12-09 11:52:46.470331943 +0000 UTC m=+1291.711090472" observedRunningTime="2025-12-09 11:52:48.691974691 +0000 UTC m=+1293.932733220" watchObservedRunningTime="2025-12-09 11:52:48.6978269 +0000 UTC m=+1293.938585419" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.706117 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltlns\" (UniqueName: \"kubernetes.io/projected/7150d0a7-b185-4942-bd70-3232d267f23a-kube-api-access-ltlns\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.706142 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.706151 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7150d0a7-b185-4942-bd70-3232d267f23a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.772163 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.780409 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-svm7p"] Dec 09 11:52:48 crc kubenswrapper[4770]: I1209 11:52:48.981820 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:52:49 crc kubenswrapper[4770]: E1209 11:52:49.312091 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod277b0a2e_73f7_4df2_9e01_d0145f26c13c.slice/crio-1674d9400cde6a620f36617354b4773df843ea2997aafea12c9d03a72568c0b7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod277b0a2e_73f7_4df2_9e01_d0145f26c13c.slice/crio-conmon-1674d9400cde6a620f36617354b4773df843ea2997aafea12c9d03a72568c0b7.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:52:49 crc kubenswrapper[4770]: I1209 11:52:49.370370 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7150d0a7-b185-4942-bd70-3232d267f23a" path="/var/lib/kubelet/pods/7150d0a7-b185-4942-bd70-3232d267f23a/volumes" Dec 09 11:52:49 crc kubenswrapper[4770]: I1209 11:52:49.688240 4770 generic.go:334] "Generic (PLEG): container finished" podID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerID="1674d9400cde6a620f36617354b4773df843ea2997aafea12c9d03a72568c0b7" exitCode=0 Dec 09 11:52:49 crc kubenswrapper[4770]: I1209 11:52:49.688393 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" event={"ID":"277b0a2e-73f7-4df2-9e01-d0145f26c13c","Type":"ContainerDied","Data":"1674d9400cde6a620f36617354b4773df843ea2997aafea12c9d03a72568c0b7"} Dec 09 11:52:49 crc kubenswrapper[4770]: I1209 11:52:49.688819 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" event={"ID":"277b0a2e-73f7-4df2-9e01-d0145f26c13c","Type":"ContainerStarted","Data":"49759d3f2b7ec95c516039c5c263a6f1b333f84550c14fb0c80ea59aab56c403"} Dec 09 11:52:50 crc kubenswrapper[4770]: I1209 11:52:50.356523 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:50 crc kubenswrapper[4770]: I1209 11:52:50.406204 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:50 crc kubenswrapper[4770]: I1209 11:52:50.698711 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" event={"ID":"277b0a2e-73f7-4df2-9e01-d0145f26c13c","Type":"ContainerStarted","Data":"64058d2e13281efc9c804ad49826b65727e1a62f615bf82bc8ad24bb102eaa4d"} Dec 09 11:52:50 crc kubenswrapper[4770]: I1209 11:52:50.698876 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:50 crc kubenswrapper[4770]: I1209 11:52:50.727194 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" podStartSLOduration=2.727168781 podStartE2EDuration="2.727168781s" podCreationTimestamp="2025-12-09 11:52:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:52:50.717038754 +0000 UTC m=+1295.957797303" watchObservedRunningTime="2025-12-09 11:52:50.727168781 +0000 UTC m=+1295.967927310" Dec 09 11:52:51 crc kubenswrapper[4770]: I1209 11:52:51.705362 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:51 crc kubenswrapper[4770]: I1209 11:52:51.751965 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 09 11:52:53 crc kubenswrapper[4770]: I1209 11:52:53.721653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerStarted","Data":"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45"} Dec 09 11:52:54 crc kubenswrapper[4770]: I1209 11:52:54.730237 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerStarted","Data":"010a70aaa353dcb251bebda226955ab3b511ce82dd40a3e21356735a99fc680e"} Dec 09 11:52:55 crc kubenswrapper[4770]: I1209 11:52:55.619055 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:55 crc kubenswrapper[4770]: I1209 11:52:55.738838 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76483913-5fec-49d0-8d19-916a024ec7fd","Type":"ContainerStarted","Data":"318e973a919a8e953fba7ee24962721c2b32ac548f3e70514be2e873a194e57c"} Dec 09 11:52:55 crc kubenswrapper[4770]: I1209 11:52:55.740319 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 11:52:56 crc kubenswrapper[4770]: I1209 11:52:56.364568 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=5.101236305 podStartE2EDuration="59.364549074s" podCreationTimestamp="2025-12-09 11:51:57 +0000 UTC" firstStartedPulling="2025-12-09 11:52:00.452800799 +0000 UTC m=+1245.693559318" lastFinishedPulling="2025-12-09 11:52:54.716113568 +0000 UTC m=+1299.956872087" observedRunningTime="2025-12-09 11:52:55.761309969 +0000 UTC m=+1301.002068488" watchObservedRunningTime="2025-12-09 11:52:56.364549074 +0000 UTC m=+1301.605307593" Dec 09 11:52:56 crc kubenswrapper[4770]: I1209 11:52:56.748088 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerStarted","Data":"9256f77b59bd0b11cf351f783c242ba7e3eb289025a5f44fd213395ee4bf575a"} Dec 09 11:52:56 crc kubenswrapper[4770]: I1209 11:52:56.750304 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerStarted","Data":"d370b0cd44dd3020056530066f02dab0b0bb51657bf84b49247c082765bd2e1d"} Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.760370 4770 generic.go:334] "Generic (PLEG): container finished" podID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerID="ce5f3ea540f088f455b54f1fb0bf39f592ea7d27d976af881d2c46b7334b8cb7" exitCode=0 Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.760471 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerDied","Data":"ce5f3ea540f088f455b54f1fb0bf39f592ea7d27d976af881d2c46b7334b8cb7"} Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.764308 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerID="ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45" exitCode=0 Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.764360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerDied","Data":"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45"} Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.768435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4" event={"ID":"f754e609-fd35-4c45-bfe8-71c659a22cdb","Type":"ContainerStarted","Data":"819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02"} Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.768687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xhnq4" Dec 09 11:52:57 crc kubenswrapper[4770]: I1209 11:52:57.812461 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xhnq4" podStartSLOduration=10.725698995 podStartE2EDuration="58.812439803s" podCreationTimestamp="2025-12-09 11:51:59 +0000 UTC" firstStartedPulling="2025-12-09 11:52:08.696541822 +0000 UTC m=+1253.937300341" lastFinishedPulling="2025-12-09 11:52:56.78328263 +0000 UTC m=+1302.024041149" observedRunningTime="2025-12-09 11:52:57.807620691 +0000 UTC m=+1303.048379220" watchObservedRunningTime="2025-12-09 11:52:57.812439803 +0000 UTC m=+1303.053198322" Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.534123 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.611024 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.611476 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7878659675-g2r4x" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="dnsmasq-dns" containerID="cri-o://67c7f2fa7c544e10472b827f869f6a0dbb8df243afedca85c04542d0ccbb42ad" gracePeriod=10 Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.920114 4770 generic.go:334] "Generic (PLEG): container finished" podID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerID="010a70aaa353dcb251bebda226955ab3b511ce82dd40a3e21356735a99fc680e" exitCode=0 Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.920217 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerDied","Data":"010a70aaa353dcb251bebda226955ab3b511ce82dd40a3e21356735a99fc680e"} Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.929829 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerStarted","Data":"5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5"} Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.929879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerStarted","Data":"37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b"} Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.930155 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.930203 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.934989 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerStarted","Data":"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b"} Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.938104 4770 generic.go:334] "Generic (PLEG): container finished" podID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerID="67c7f2fa7c544e10472b827f869f6a0dbb8df243afedca85c04542d0ccbb42ad" exitCode=0 Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.938186 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-g2r4x" event={"ID":"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7","Type":"ContainerDied","Data":"67c7f2fa7c544e10472b827f869f6a0dbb8df243afedca85c04542d0ccbb42ad"} Dec 09 11:52:58 crc kubenswrapper[4770]: I1209 11:52:58.988534 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.643612526 podStartE2EDuration="1m8.988512742s" podCreationTimestamp="2025-12-09 11:51:50 +0000 UTC" firstStartedPulling="2025-12-09 11:51:54.502171835 +0000 UTC m=+1239.742930354" lastFinishedPulling="2025-12-09 11:52:52.847072051 +0000 UTC m=+1298.087830570" observedRunningTime="2025-12-09 11:52:58.975453651 +0000 UTC m=+1304.216212170" watchObservedRunningTime="2025-12-09 11:52:58.988512742 +0000 UTC m=+1304.229271261" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.000950 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gct95" podStartSLOduration=12.023820278 podStartE2EDuration="1m0.000923567s" podCreationTimestamp="2025-12-09 11:51:59 +0000 UTC" firstStartedPulling="2025-12-09 11:52:08.804974991 +0000 UTC m=+1254.045733520" lastFinishedPulling="2025-12-09 11:52:56.78207829 +0000 UTC m=+1302.022836809" observedRunningTime="2025-12-09 11:52:58.9982969 +0000 UTC m=+1304.239055419" watchObservedRunningTime="2025-12-09 11:52:59.000923567 +0000 UTC m=+1304.241682106" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.222245 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.291156 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config\") pod \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.291380 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkj96\" (UniqueName: \"kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96\") pod \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.291414 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc\") pod \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.291472 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb\") pod \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\" (UID: \"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7\") " Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.296630 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96" (OuterVolumeSpecName: "kube-api-access-tkj96") pod "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" (UID: "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7"). InnerVolumeSpecName "kube-api-access-tkj96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.338529 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" (UID: "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.344295 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config" (OuterVolumeSpecName: "config") pod "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" (UID: "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.349210 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" (UID: "81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.484941 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkj96\" (UniqueName: \"kubernetes.io/projected/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-kube-api-access-tkj96\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.484979 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.484992 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.485036 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:52:59 crc kubenswrapper[4770]: E1209 11:52:59.682917 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81bf1b9e_0c6e_43d8_81df_a16fb7aa59a7.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.947797 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-g2r4x" event={"ID":"81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7","Type":"ContainerDied","Data":"98393f4f7c1f9414a8e905e6561d3fdffa792dee7064ef74bf062533fcdb7091"} Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.947843 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-g2r4x" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.947864 4770 scope.go:117] "RemoveContainer" containerID="67c7f2fa7c544e10472b827f869f6a0dbb8df243afedca85c04542d0ccbb42ad" Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.950097 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerStarted","Data":"2946bef94b26e9eb5227b621b0591908bb8f5db0d79016f7e9097a651629b7fc"} Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.952084 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f352a11c-1887-4736-98cb-4eccf3086e97","Type":"ContainerStarted","Data":"a7dc5c4c948b7e59408c0e3cf9bcfbf9cd1717522c72afefd9d893df384cec6a"} Dec 09 11:52:59 crc kubenswrapper[4770]: I1209 11:52:59.982284 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371969.872515 podStartE2EDuration="1m6.982261728s" podCreationTimestamp="2025-12-09 11:51:53 +0000 UTC" firstStartedPulling="2025-12-09 11:51:57.028342534 +0000 UTC m=+1242.269101053" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:52:59.973696541 +0000 UTC m=+1305.214455060" watchObservedRunningTime="2025-12-09 11:52:59.982261728 +0000 UTC m=+1305.223020247" Dec 09 11:53:00 crc kubenswrapper[4770]: I1209 11:53:00.004143 4770 scope.go:117] "RemoveContainer" containerID="63b6b28586679902472eef833c7b0c5e21e38bbedc017c3ddc0df2decab51165" Dec 09 11:53:00 crc kubenswrapper[4770]: I1209 11:53:00.004780 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=4.322583078 podStartE2EDuration="1m6.004759988s" podCreationTimestamp="2025-12-09 11:51:54 +0000 UTC" firstStartedPulling="2025-12-09 11:51:57.45650382 +0000 UTC m=+1242.697262339" lastFinishedPulling="2025-12-09 11:52:59.13868073 +0000 UTC m=+1304.379439249" observedRunningTime="2025-12-09 11:53:00.001328981 +0000 UTC m=+1305.242087500" watchObservedRunningTime="2025-12-09 11:53:00.004759988 +0000 UTC m=+1305.245518507" Dec 09 11:53:00 crc kubenswrapper[4770]: I1209 11:53:00.049828 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:53:00 crc kubenswrapper[4770]: I1209 11:53:00.056645 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7878659675-g2r4x"] Dec 09 11:53:01 crc kubenswrapper[4770]: I1209 11:53:01.149876 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 09 11:53:01 crc kubenswrapper[4770]: I1209 11:53:01.354671 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" path="/var/lib/kubelet/pods/81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7/volumes" Dec 09 11:53:03 crc kubenswrapper[4770]: I1209 11:53:03.170511 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 11:53:03 crc kubenswrapper[4770]: I1209 11:53:03.170802 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 11:53:03 crc kubenswrapper[4770]: I1209 11:53:03.256099 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 09 11:53:03 crc kubenswrapper[4770]: I1209 11:53:03.990120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerStarted","Data":"3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b"} Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.014258 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.70673294 podStartE2EDuration="1m5.014239616s" podCreationTimestamp="2025-12-09 11:51:59 +0000 UTC" firstStartedPulling="2025-12-09 11:52:02.724148457 +0000 UTC m=+1247.964906976" lastFinishedPulling="2025-12-09 11:53:03.031655133 +0000 UTC m=+1308.272413652" observedRunningTime="2025-12-09 11:53:04.01360534 +0000 UTC m=+1309.254363879" watchObservedRunningTime="2025-12-09 11:53:04.014239616 +0000 UTC m=+1309.254998135" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.067058 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.071484 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.576281 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-fmjv6"] Dec 09 11:53:04 crc kubenswrapper[4770]: E1209 11:53:04.576736 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="init" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.576751 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="init" Dec 09 11:53:04 crc kubenswrapper[4770]: E1209 11:53:04.576779 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="dnsmasq-dns" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.576787 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="dnsmasq-dns" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.577003 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="81bf1b9e-0c6e-43d8-81df-a16fb7aa59a7" containerName="dnsmasq-dns" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.577682 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.603468 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9545-account-create-update-4vn85"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.605888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.608330 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.627316 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fmjv6"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.649713 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9545-account-create-update-4vn85"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.674580 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb8wg\" (UniqueName: \"kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.674641 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.676386 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-xx2q9"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.677533 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.697629 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xx2q9"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.748037 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6c0f-account-create-update-sstlm"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.749548 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.751640 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.774385 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c0f-account-create-update-sstlm"] Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.859615 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb8wg\" (UniqueName: \"kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.859666 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.859753 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.859850 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm74b\" (UniqueName: \"kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.859975 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.860021 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2ts\" (UniqueName: \"kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.860055 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.860127 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5ztf\" (UniqueName: \"kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.861509 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.895086 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb8wg\" (UniqueName: \"kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg\") pod \"keystone-db-create-fmjv6\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.913342 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.923023 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.923054 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962367 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2ts\" (UniqueName: \"kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962434 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962492 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5ztf\" (UniqueName: \"kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962534 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962593 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm74b\" (UniqueName: \"kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.962642 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.963418 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.963456 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.964015 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.995600 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2ts\" (UniqueName: \"kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts\") pod \"placement-6c0f-account-create-update-sstlm\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.996432 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5ztf\" (UniqueName: \"kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf\") pod \"placement-db-create-xx2q9\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:04 crc kubenswrapper[4770]: I1209 11:53:04.997144 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.005040 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm74b\" (UniqueName: \"kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b\") pod \"keystone-9545-account-create-update-4vn85\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.178633 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.234021 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.247965 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.453935 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.509723 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fmjv6"] Dec 09 11:53:05 crc kubenswrapper[4770]: W1209 11:53:05.514526 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18ff5c57_41ce_43a3_b65e_7d52dd45ee3f.slice/crio-9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d WatchSource:0}: Error finding container 9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d: Status 404 returned error can't find the container with id 9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.793097 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xx2q9"] Dec 09 11:53:05 crc kubenswrapper[4770]: W1209 11:53:05.796135 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c72d3ee_7e15_4402_940d_a4524531a11d.slice/crio-9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1 WatchSource:0}: Error finding container 9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1: Status 404 returned error can't find the container with id 9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1 Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.854985 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c0f-account-create-update-sstlm"] Dec 09 11:53:05 crc kubenswrapper[4770]: W1209 11:53:05.858164 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f916059_5f99_4cb8_aac7_0f1c65f0eb1e.slice/crio-a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3 WatchSource:0}: Error finding container a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3: Status 404 returned error can't find the container with id a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3 Dec 09 11:53:05 crc kubenswrapper[4770]: I1209 11:53:05.931863 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9545-account-create-update-4vn85"] Dec 09 11:53:05 crc kubenswrapper[4770]: W1209 11:53:05.934923 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5dc7848_9e9b_43e2_8800_b7c6678c1eab.slice/crio-9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39 WatchSource:0}: Error finding container 9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39: Status 404 returned error can't find the container with id 9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39 Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.045172 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx2q9" event={"ID":"3c72d3ee-7e15-4402-940d-a4524531a11d","Type":"ContainerStarted","Data":"bf9a900264b568b10dc3739abfa123854754954e74d507968ba4ce9a0b201378"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.045237 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx2q9" event={"ID":"3c72d3ee-7e15-4402-940d-a4524531a11d","Type":"ContainerStarted","Data":"9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.048038 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9545-account-create-update-4vn85" event={"ID":"a5dc7848-9e9b-43e2-8800-b7c6678c1eab","Type":"ContainerStarted","Data":"9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.056719 4770 generic.go:334] "Generic (PLEG): container finished" podID="18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" containerID="aaa58da97bffc14fa7574c9cd57935763dac5dc795f51d4a296070911eb74b14" exitCode=0 Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.056827 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fmjv6" event={"ID":"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f","Type":"ContainerDied","Data":"aaa58da97bffc14fa7574c9cd57935763dac5dc795f51d4a296070911eb74b14"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.056858 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fmjv6" event={"ID":"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f","Type":"ContainerStarted","Data":"9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.062090 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c0f-account-create-update-sstlm" event={"ID":"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e","Type":"ContainerStarted","Data":"a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3"} Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.064403 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-xx2q9" podStartSLOduration=2.064385306 podStartE2EDuration="2.064385306s" podCreationTimestamp="2025-12-09 11:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:06.060003785 +0000 UTC m=+1311.300762294" watchObservedRunningTime="2025-12-09 11:53:06.064385306 +0000 UTC m=+1311.305143825" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.067017 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.184460 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.356772 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6x5x2"] Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.358461 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.365894 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6x5x2"] Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.454835 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-078e-account-create-update-d6hrg"] Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.456519 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.458448 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.467386 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-078e-account-create-update-d6hrg"] Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.490959 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.491123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wlcd\" (UniqueName: \"kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.593087 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljw57\" (UniqueName: \"kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.593140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.593413 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wlcd\" (UniqueName: \"kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.593923 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.594725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.615275 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wlcd\" (UniqueName: \"kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd\") pod \"glance-db-create-6x5x2\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.682771 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.695240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljw57\" (UniqueName: \"kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.695307 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.698199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.715953 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljw57\" (UniqueName: \"kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57\") pod \"glance-078e-account-create-update-d6hrg\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:06 crc kubenswrapper[4770]: I1209 11:53:06.774264 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.086396 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx2q9" event={"ID":"3c72d3ee-7e15-4402-940d-a4524531a11d","Type":"ContainerDied","Data":"bf9a900264b568b10dc3739abfa123854754954e74d507968ba4ce9a0b201378"} Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.086649 4770 generic.go:334] "Generic (PLEG): container finished" podID="3c72d3ee-7e15-4402-940d-a4524531a11d" containerID="bf9a900264b568b10dc3739abfa123854754954e74d507968ba4ce9a0b201378" exitCode=0 Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.091389 4770 generic.go:334] "Generic (PLEG): container finished" podID="a5dc7848-9e9b-43e2-8800-b7c6678c1eab" containerID="74e47b78fc3704d9f254adeaf8bbb3db0c5dd90dcb22750ff79aaa07926d8e44" exitCode=0 Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.091465 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9545-account-create-update-4vn85" event={"ID":"a5dc7848-9e9b-43e2-8800-b7c6678c1eab","Type":"ContainerDied","Data":"74e47b78fc3704d9f254adeaf8bbb3db0c5dd90dcb22750ff79aaa07926d8e44"} Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.093306 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" containerID="ce7b985dd5d6ad69d34a28139df573f6e591655833d1d03998c27319b1dd79e8" exitCode=0 Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.093363 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c0f-account-create-update-sstlm" event={"ID":"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e","Type":"ContainerDied","Data":"ce7b985dd5d6ad69d34a28139df573f6e591655833d1d03998c27319b1dd79e8"} Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.127256 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.162548 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6x5x2"] Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.286807 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-078e-account-create-update-d6hrg"] Dec 09 11:53:07 crc kubenswrapper[4770]: W1209 11:53:07.287610 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1edf704f_d8b6_4205_a818_bbcae6131f58.slice/crio-166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267 WatchSource:0}: Error finding container 166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267: Status 404 returned error can't find the container with id 166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267 Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.442517 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.622834 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts\") pod \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.622876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb8wg\" (UniqueName: \"kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg\") pod \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\" (UID: \"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f\") " Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.623339 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" (UID: "18ff5c57-41ce-43a3-b65e-7d52dd45ee3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.628945 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg" (OuterVolumeSpecName: "kube-api-access-wb8wg") pod "18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" (UID: "18ff5c57-41ce-43a3-b65e-7d52dd45ee3f"). InnerVolumeSpecName "kube-api-access-wb8wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.724609 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:07 crc kubenswrapper[4770]: I1209 11:53:07.724646 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb8wg\" (UniqueName: \"kubernetes.io/projected/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f-kube-api-access-wb8wg\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.121597 4770 generic.go:334] "Generic (PLEG): container finished" podID="bffb1e84-2444-47b4-9152-1920771323c6" containerID="377a052b321d83d9e8bff8487f5444db77436defd96fd805994e7fdf82aecba9" exitCode=0 Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.121775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6x5x2" event={"ID":"bffb1e84-2444-47b4-9152-1920771323c6","Type":"ContainerDied","Data":"377a052b321d83d9e8bff8487f5444db77436defd96fd805994e7fdf82aecba9"} Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.121975 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6x5x2" event={"ID":"bffb1e84-2444-47b4-9152-1920771323c6","Type":"ContainerStarted","Data":"634ebd13dcc9651da720a7adc93bcc3b124935ca4b023ab29023183e06895f07"} Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.124877 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fmjv6" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.124908 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fmjv6" event={"ID":"18ff5c57-41ce-43a3-b65e-7d52dd45ee3f","Type":"ContainerDied","Data":"9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d"} Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.124973 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ca333029f3a4ff6d5cc94aba37ab3afe32fb379099329f505bfe5e08992bd1d" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.126792 4770 generic.go:334] "Generic (PLEG): container finished" podID="1edf704f-d8b6-4205-a818-bbcae6131f58" containerID="7d29da55d749c371a137286ed3944f43b67a190a65d6103836dab022c75fbc05" exitCode=0 Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.127022 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-078e-account-create-update-d6hrg" event={"ID":"1edf704f-d8b6-4205-a818-bbcae6131f58","Type":"ContainerDied","Data":"7d29da55d749c371a137286ed3944f43b67a190a65d6103836dab022c75fbc05"} Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.127043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-078e-account-create-update-d6hrg" event={"ID":"1edf704f-d8b6-4205-a818-bbcae6131f58","Type":"ContainerStarted","Data":"166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267"} Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.144261 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.637865 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:53:08 crc kubenswrapper[4770]: E1209 11:53:08.638273 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" containerName="mariadb-database-create" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.638286 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" containerName="mariadb-database-create" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.638458 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" containerName="mariadb-database-create" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.639333 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.656982 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.746747 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.747143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wvx\" (UniqueName: \"kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.747173 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.747217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.747257 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.793914 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.849234 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.849287 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wvx\" (UniqueName: \"kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.849319 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.849357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.849390 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.853907 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.854570 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.855231 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.855765 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.882055 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wvx\" (UniqueName: \"kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx\") pod \"dnsmasq-dns-67fdf7998c-klzb9\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.950448 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts\") pod \"3c72d3ee-7e15-4402-940d-a4524531a11d\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.950648 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5ztf\" (UniqueName: \"kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf\") pod \"3c72d3ee-7e15-4402-940d-a4524531a11d\" (UID: \"3c72d3ee-7e15-4402-940d-a4524531a11d\") " Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.950933 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c72d3ee-7e15-4402-940d-a4524531a11d" (UID: "3c72d3ee-7e15-4402-940d-a4524531a11d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.951133 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c72d3ee-7e15-4402-940d-a4524531a11d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:08 crc kubenswrapper[4770]: I1209 11:53:08.954405 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf" (OuterVolumeSpecName: "kube-api-access-v5ztf") pod "3c72d3ee-7e15-4402-940d-a4524531a11d" (UID: "3c72d3ee-7e15-4402-940d-a4524531a11d"). InnerVolumeSpecName "kube-api-access-v5ztf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.032193 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.041122 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.052973 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5ztf\" (UniqueName: \"kubernetes.io/projected/3c72d3ee-7e15-4402-940d-a4524531a11d-kube-api-access-v5ztf\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.087490 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.193574 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts\") pod \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.193655 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts\") pod \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.193739 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm74b\" (UniqueName: \"kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b\") pod \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\" (UID: \"a5dc7848-9e9b-43e2-8800-b7c6678c1eab\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.193851 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d2ts\" (UniqueName: \"kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts\") pod \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\" (UID: \"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.195019 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5dc7848-9e9b-43e2-8800-b7c6678c1eab" (UID: "a5dc7848-9e9b-43e2-8800-b7c6678c1eab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.195502 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" (UID: "6f916059-5f99-4cb8-aac7-0f1c65f0eb1e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.203254 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts" (OuterVolumeSpecName: "kube-api-access-6d2ts") pod "6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" (UID: "6f916059-5f99-4cb8-aac7-0f1c65f0eb1e"). InnerVolumeSpecName "kube-api-access-6d2ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.219461 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b" (OuterVolumeSpecName: "kube-api-access-wm74b") pod "a5dc7848-9e9b-43e2-8800-b7c6678c1eab" (UID: "a5dc7848-9e9b-43e2-8800-b7c6678c1eab"). InnerVolumeSpecName "kube-api-access-wm74b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.235073 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9545-account-create-update-4vn85" event={"ID":"a5dc7848-9e9b-43e2-8800-b7c6678c1eab","Type":"ContainerDied","Data":"9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39"} Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.235126 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fca8b71a089f9589cf57d009f387b4a744a67cbba61f9c3bf62b7fcee3a8e39" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.235207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9545-account-create-update-4vn85" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.274162 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c0f-account-create-update-sstlm" event={"ID":"6f916059-5f99-4cb8-aac7-0f1c65f0eb1e","Type":"ContainerDied","Data":"a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3"} Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.274565 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55ca5130fe8c54d4b4deeca6630fb0b54513392ea28f2bb31da9552b9e132d3" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.274651 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c0f-account-create-update-sstlm" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.283535 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx2q9" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.285686 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx2q9" event={"ID":"3c72d3ee-7e15-4402-940d-a4524531a11d","Type":"ContainerDied","Data":"9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1"} Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.285734 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dd123ccaaf50e27d92ef3600bc18260e8822194ff16e1c94ee5f1c57a222dd1" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.296838 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.296868 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.296877 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm74b\" (UniqueName: \"kubernetes.io/projected/a5dc7848-9e9b-43e2-8800-b7c6678c1eab-kube-api-access-wm74b\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.296887 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d2ts\" (UniqueName: \"kubernetes.io/projected/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e-kube-api-access-6d2ts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.637578 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.644894 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.711195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts\") pod \"bffb1e84-2444-47b4-9152-1920771323c6\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.711260 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts\") pod \"1edf704f-d8b6-4205-a818-bbcae6131f58\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.711343 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljw57\" (UniqueName: \"kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57\") pod \"1edf704f-d8b6-4205-a818-bbcae6131f58\" (UID: \"1edf704f-d8b6-4205-a818-bbcae6131f58\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.711498 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wlcd\" (UniqueName: \"kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd\") pod \"bffb1e84-2444-47b4-9152-1920771323c6\" (UID: \"bffb1e84-2444-47b4-9152-1920771323c6\") " Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.712284 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1edf704f-d8b6-4205-a818-bbcae6131f58" (UID: "1edf704f-d8b6-4205-a818-bbcae6131f58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.712423 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bffb1e84-2444-47b4-9152-1920771323c6" (UID: "bffb1e84-2444-47b4-9152-1920771323c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.719619 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57" (OuterVolumeSpecName: "kube-api-access-ljw57") pod "1edf704f-d8b6-4205-a818-bbcae6131f58" (UID: "1edf704f-d8b6-4205-a818-bbcae6131f58"). InnerVolumeSpecName "kube-api-access-ljw57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.720216 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd" (OuterVolumeSpecName: "kube-api-access-8wlcd") pod "bffb1e84-2444-47b4-9152-1920771323c6" (UID: "bffb1e84-2444-47b4-9152-1920771323c6"). InnerVolumeSpecName "kube-api-access-8wlcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753482 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.753848 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bffb1e84-2444-47b4-9152-1920771323c6" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753866 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bffb1e84-2444-47b4-9152-1920771323c6" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.753877 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c72d3ee-7e15-4402-940d-a4524531a11d" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753883 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c72d3ee-7e15-4402-940d-a4524531a11d" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.753916 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753924 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.753936 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5dc7848-9e9b-43e2-8800-b7c6678c1eab" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753942 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5dc7848-9e9b-43e2-8800-b7c6678c1eab" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.753957 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edf704f-d8b6-4205-a818-bbcae6131f58" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.753963 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edf704f-d8b6-4205-a818-bbcae6131f58" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.754127 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bffb1e84-2444-47b4-9152-1920771323c6" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.754145 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.754154 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5dc7848-9e9b-43e2-8800-b7c6678c1eab" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.754165 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edf704f-d8b6-4205-a818-bbcae6131f58" containerName="mariadb-account-create-update" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.754174 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c72d3ee-7e15-4402-940d-a4524531a11d" containerName="mariadb-database-create" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.759093 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.766408 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.766516 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-zms65" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.766655 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.766745 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.777664 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.803490 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813219 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxcsl\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813262 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813415 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813451 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813501 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813654 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wlcd\" (UniqueName: \"kubernetes.io/projected/bffb1e84-2444-47b4-9152-1920771323c6-kube-api-access-8wlcd\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813677 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bffb1e84-2444-47b4-9152-1920771323c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813690 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf704f-d8b6-4205-a818-bbcae6131f58-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.813703 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljw57\" (UniqueName: \"kubernetes.io/projected/1edf704f-d8b6-4205-a818-bbcae6131f58-kube-api-access-ljw57\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.914709 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxcsl\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.914755 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.914822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.914843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.914871 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.915097 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.915119 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:09 crc kubenswrapper[4770]: E1209 11:53:09.915168 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:10.415148509 +0000 UTC m=+1315.655907028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.915297 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.915466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.915946 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.935472 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxcsl\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:09 crc kubenswrapper[4770]: I1209 11:53:09.942255 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.291389 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" event={"ID":"ab1ee405-e664-46bc-a816-1526373ada15","Type":"ContainerStarted","Data":"433da4e8899de70be5ce6d63d2f0c0b86acaa6ff72561be5438920e6e91f87a1"} Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.293242 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6x5x2" event={"ID":"bffb1e84-2444-47b4-9152-1920771323c6","Type":"ContainerDied","Data":"634ebd13dcc9651da720a7adc93bcc3b124935ca4b023ab29023183e06895f07"} Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.293269 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634ebd13dcc9651da720a7adc93bcc3b124935ca4b023ab29023183e06895f07" Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.293294 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6x5x2" Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.294778 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-078e-account-create-update-d6hrg" event={"ID":"1edf704f-d8b6-4205-a818-bbcae6131f58","Type":"ContainerDied","Data":"166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267"} Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.294804 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="166b4710f5f9a959fe33f78824125501ce93b9f63657470f3e31864d14f46267" Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.294833 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-078e-account-create-update-d6hrg" Dec 09 11:53:10 crc kubenswrapper[4770]: I1209 11:53:10.479309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:10 crc kubenswrapper[4770]: E1209 11:53:10.479533 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:10 crc kubenswrapper[4770]: E1209 11:53:10.479552 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:10 crc kubenswrapper[4770]: E1209 11:53:10.479610 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:11.47959152 +0000 UTC m=+1316.720350039 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.113143 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.288859 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.290561 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.293071 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.300199 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.300374 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8tqc6" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.300447 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.327242 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab1ee405-e664-46bc-a816-1526373ada15" containerID="d501d6b22d091e0904ae87a07713eca3b978c87e116a5d4755b1a061f6aa2702" exitCode=0 Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.327294 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" event={"ID":"ab1ee405-e664-46bc-a816-1526373ada15","Type":"ContainerDied","Data":"d501d6b22d091e0904ae87a07713eca3b978c87e116a5d4755b1a061f6aa2702"} Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.329276 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.404661 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405226 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405415 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405450 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhpbn\" (UniqueName: \"kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.405595 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507355 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507390 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507420 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507451 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507489 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507513 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhpbn\" (UniqueName: \"kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507567 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: E1209 11:53:11.507592 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:11 crc kubenswrapper[4770]: E1209 11:53:11.507621 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:11 crc kubenswrapper[4770]: E1209 11:53:11.507674 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:13.507657315 +0000 UTC m=+1318.748415824 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.507878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.508828 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.509221 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.511875 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.512562 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.522708 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.527030 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhpbn\" (UniqueName: \"kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn\") pod \"ovn-northd-0\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.626493 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-pxdc5"] Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.627852 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.633798 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.634636 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fvqzp" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.637081 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pxdc5"] Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.662135 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.712012 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whvds\" (UniqueName: \"kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.712096 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.712134 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:11 crc kubenswrapper[4770]: I1209 11:53:11.712296 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.087128 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whvds\" (UniqueName: \"kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.087541 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.087639 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.087969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.102228 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.104381 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.112734 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.135840 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whvds\" (UniqueName: \"kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds\") pod \"glance-db-sync-pxdc5\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.245161 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pxdc5" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.411010 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" event={"ID":"ab1ee405-e664-46bc-a816-1526373ada15","Type":"ContainerStarted","Data":"dcfabf3d3df79b39f1a1ad99c77d0fba948ca668b9b09caaf8c8a7c3db3857b4"} Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.413203 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.444843 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" podStartSLOduration=4.444816086 podStartE2EDuration="4.444816086s" podCreationTimestamp="2025-12-09 11:53:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:12.440394404 +0000 UTC m=+1317.681152923" watchObservedRunningTime="2025-12-09 11:53:12.444816086 +0000 UTC m=+1317.685574615" Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.704639 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:53:12 crc kubenswrapper[4770]: W1209 11:53:12.705675 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c7c3022_77d1_4055_8d10_6c7a474a9833.slice/crio-962fa3b338d0c548a486cb5aa72063b0086c5248ec02a5ce01349ac200afd6a4 WatchSource:0}: Error finding container 962fa3b338d0c548a486cb5aa72063b0086c5248ec02a5ce01349ac200afd6a4: Status 404 returned error can't find the container with id 962fa3b338d0c548a486cb5aa72063b0086c5248ec02a5ce01349ac200afd6a4 Dec 09 11:53:12 crc kubenswrapper[4770]: I1209 11:53:12.947880 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pxdc5"] Dec 09 11:53:12 crc kubenswrapper[4770]: W1209 11:53:12.949726 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c091c39_7a67_414c_9cbd_d323073c87be.slice/crio-c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f WatchSource:0}: Error finding container c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f: Status 404 returned error can't find the container with id c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.422462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pxdc5" event={"ID":"0c091c39-7a67-414c-9cbd-d323073c87be","Type":"ContainerStarted","Data":"c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f"} Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.424499 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerStarted","Data":"962fa3b338d0c548a486cb5aa72063b0086c5248ec02a5ce01349ac200afd6a4"} Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.529329 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:13 crc kubenswrapper[4770]: E1209 11:53:13.529892 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:13 crc kubenswrapper[4770]: E1209 11:53:13.529925 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:13 crc kubenswrapper[4770]: E1209 11:53:13.529965 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:17.529950859 +0000 UTC m=+1322.770709368 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.628791 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2n7xh"] Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.631582 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.635348 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.636585 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.646179 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.710572 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2n7xh"] Dec 09 11:53:13 crc kubenswrapper[4770]: E1209 11:53:13.711043 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-jdp5n ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-jdp5n ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-2n7xh" podUID="a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.718998 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-v7q9j"] Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.727518 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-v7q9j"] Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.727615 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.736080 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2n7xh"] Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.737683 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.737771 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdp5n\" (UniqueName: \"kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.738123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.738345 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.738366 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.738563 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.738628 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840545 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840629 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-625x5\" (UniqueName: \"kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840675 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840706 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840733 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840771 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840822 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840851 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdp5n\" (UniqueName: \"kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840964 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.840998 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.841070 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.841097 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.841140 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.841680 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.842394 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.843761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.850661 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.850712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.865441 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.866617 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdp5n\" (UniqueName: \"kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n\") pod \"swift-ring-rebalance-2n7xh\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942498 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942565 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942611 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942721 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.942950 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.943014 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-625x5\" (UniqueName: \"kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.943424 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.943783 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.945930 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.952503 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.956871 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.958377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:13 crc kubenswrapper[4770]: I1209 11:53:13.963159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-625x5\" (UniqueName: \"kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5\") pod \"swift-ring-rebalance-v7q9j\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.055326 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.434262 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.446058 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554573 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdp5n\" (UniqueName: \"kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554709 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554761 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.554832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.556166 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices\") pod \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\" (UID: \"a53ccfaf-41a3-4e3b-8225-7ad24d49d80b\") " Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.557816 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts" (OuterVolumeSpecName: "scripts") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.558112 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.559095 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.561949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n" (OuterVolumeSpecName: "kube-api-access-jdp5n") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "kube-api-access-jdp5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.561958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.561982 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.563577 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" (UID: "a53ccfaf-41a3-4e3b-8225-7ad24d49d80b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.632636 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-v7q9j"] Dec 09 11:53:14 crc kubenswrapper[4770]: W1209 11:53:14.656630 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fdfdc9d_aa13_4fdb_8e1f_c54801e47867.slice/crio-eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3 WatchSource:0}: Error finding container eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3: Status 404 returned error can't find the container with id eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3 Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659093 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659225 4770 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659242 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659254 4770 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659267 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659278 4770 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:14 crc kubenswrapper[4770]: I1209 11:53:14.659290 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdp5n\" (UniqueName: \"kubernetes.io/projected/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b-kube-api-access-jdp5n\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.448256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerStarted","Data":"60b3561b63896f5bc9807fe5e6de66a0dadd48252dfe9a1b9a2f39c48715e956"} Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.448616 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.448673 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerStarted","Data":"867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870"} Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.452166 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2n7xh" Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.452174 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-v7q9j" event={"ID":"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867","Type":"ContainerStarted","Data":"eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3"} Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.473993 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.967154783 podStartE2EDuration="4.473973438s" podCreationTimestamp="2025-12-09 11:53:11 +0000 UTC" firstStartedPulling="2025-12-09 11:53:12.708615725 +0000 UTC m=+1317.949374244" lastFinishedPulling="2025-12-09 11:53:14.21543438 +0000 UTC m=+1319.456192899" observedRunningTime="2025-12-09 11:53:15.464941929 +0000 UTC m=+1320.705700468" watchObservedRunningTime="2025-12-09 11:53:15.473973438 +0000 UTC m=+1320.714731957" Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.504092 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2n7xh"] Dec 09 11:53:15 crc kubenswrapper[4770]: I1209 11:53:15.512636 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-2n7xh"] Dec 09 11:53:17 crc kubenswrapper[4770]: I1209 11:53:17.353816 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53ccfaf-41a3-4e3b-8225-7ad24d49d80b" path="/var/lib/kubelet/pods/a53ccfaf-41a3-4e3b-8225-7ad24d49d80b/volumes" Dec 09 11:53:17 crc kubenswrapper[4770]: I1209 11:53:17.631342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:17 crc kubenswrapper[4770]: E1209 11:53:17.631592 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:17 crc kubenswrapper[4770]: E1209 11:53:17.631621 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:17 crc kubenswrapper[4770]: E1209 11:53:17.631692 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:25.631674276 +0000 UTC m=+1330.872432795 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:19 crc kubenswrapper[4770]: I1209 11:53:19.090132 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:53:19 crc kubenswrapper[4770]: I1209 11:53:19.205840 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:53:19 crc kubenswrapper[4770]: I1209 11:53:19.206881 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" containerID="cri-o://64058d2e13281efc9c804ad49826b65727e1a62f615bf82bc8ad24bb102eaa4d" gracePeriod=10 Dec 09 11:53:19 crc kubenswrapper[4770]: I1209 11:53:19.501724 4770 generic.go:334] "Generic (PLEG): container finished" podID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerID="64058d2e13281efc9c804ad49826b65727e1a62f615bf82bc8ad24bb102eaa4d" exitCode=0 Dec 09 11:53:19 crc kubenswrapper[4770]: I1209 11:53:19.501778 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" event={"ID":"277b0a2e-73f7-4df2-9e01-d0145f26c13c","Type":"ContainerDied","Data":"64058d2e13281efc9c804ad49826b65727e1a62f615bf82bc8ad24bb102eaa4d"} Dec 09 11:53:23 crc kubenswrapper[4770]: I1209 11:53:23.533006 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Dec 09 11:53:25 crc kubenswrapper[4770]: I1209 11:53:25.681285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:25 crc kubenswrapper[4770]: E1209 11:53:25.681544 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:25 crc kubenswrapper[4770]: E1209 11:53:25.681582 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:25 crc kubenswrapper[4770]: E1209 11:53:25.681644 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:53:41.681622734 +0000 UTC m=+1346.922381253 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:26 crc kubenswrapper[4770]: I1209 11:53:26.718075 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 09 11:53:28 crc kubenswrapper[4770]: I1209 11:53:28.533054 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Dec 09 11:53:29 crc kubenswrapper[4770]: I1209 11:53:29.595030 4770 generic.go:334] "Generic (PLEG): container finished" podID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerID="9256f77b59bd0b11cf351f783c242ba7e3eb289025a5f44fd213395ee4bf575a" exitCode=0 Dec 09 11:53:29 crc kubenswrapper[4770]: I1209 11:53:29.595134 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerDied","Data":"9256f77b59bd0b11cf351f783c242ba7e3eb289025a5f44fd213395ee4bf575a"} Dec 09 11:53:29 crc kubenswrapper[4770]: I1209 11:53:29.600479 4770 generic.go:334] "Generic (PLEG): container finished" podID="03fef9de-3f78-48b3-9079-9dc87184f803" containerID="d370b0cd44dd3020056530066f02dab0b0bb51657bf84b49247c082765bd2e1d" exitCode=0 Dec 09 11:53:29 crc kubenswrapper[4770]: I1209 11:53:29.600552 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerDied","Data":"d370b0cd44dd3020056530066f02dab0b0bb51657bf84b49247c082765bd2e1d"} Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.019577 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" probeResult="failure" output=< Dec 09 11:53:30 crc kubenswrapper[4770]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 11:53:30 crc kubenswrapper[4770]: > Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.259803 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.264583 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.503941 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xhnq4-config-8rqdb"] Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.505439 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.507421 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.518979 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhnq4-config-8rqdb"] Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686109 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686556 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686638 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686705 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29bjp\" (UniqueName: \"kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686751 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.686819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789242 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789332 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29bjp\" (UniqueName: \"kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789428 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789455 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.789611 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.790145 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.791451 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.810940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29bjp\" (UniqueName: \"kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp\") pod \"ovn-controller-xhnq4-config-8rqdb\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: I1209 11:53:30.843855 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:30 crc kubenswrapper[4770]: E1209 11:53:30.943597 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f" Dec 09 11:53:30 crc kubenswrapper[4770]: E1209 11:53:30.943792 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whvds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-pxdc5_openstack(0c091c39-7a67-414c-9cbd-d323073c87be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:53:30 crc kubenswrapper[4770]: E1209 11:53:30.945014 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-pxdc5" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" Dec 09 11:53:31 crc kubenswrapper[4770]: E1209 11:53:31.615961 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f\\\"\"" pod="openstack/glance-db-sync-pxdc5" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" Dec 09 11:53:32 crc kubenswrapper[4770]: E1209 11:53:32.721232 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:ac7fefe1c93839c7ccb2aaa0a18751df0e9f64a36a3b4cc1b81d82d7774b8b45" Dec 09 11:53:32 crc kubenswrapper[4770]: E1209 11:53:32.721484 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:swift-ring-rebalance,Image:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:ac7fefe1c93839c7ccb2aaa0a18751df0e9f64a36a3b4cc1b81d82d7774b8b45,Command:[/usr/local/bin/swift-ring-tool all],Args:[],WorkingDir:/etc/swift,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CM_NAME,Value:swift-ring-files,ValueFrom:nil,},EnvVar{Name:NAMESPACE,Value:openstack,ValueFrom:nil,},EnvVar{Name:OWNER_APIVERSION,Value:swift.openstack.org/v1beta1,ValueFrom:nil,},EnvVar{Name:OWNER_KIND,Value:SwiftRing,ValueFrom:nil,},EnvVar{Name:OWNER_NAME,Value:swift-ring,ValueFrom:nil,},EnvVar{Name:OWNER_UID,Value:17a38ab9-5af0-4a7a-ade5-836d3519da75,ValueFrom:nil,},EnvVar{Name:SWIFT_MIN_PART_HOURS,Value:1,ValueFrom:nil,},EnvVar{Name:SWIFT_PART_POWER,Value:10,ValueFrom:nil,},EnvVar{Name:SWIFT_REPLICAS,Value:1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/swift-ring-tool,SubPath:swift-ring-tool,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:swiftconf,ReadOnly:true,MountPath:/etc/swift/swift.conf,SubPath:swift.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ring-data-devices,ReadOnly:true,MountPath:/var/lib/config-data/ring-devices,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dispersionconf,ReadOnly:true,MountPath:/etc/swift/dispersion.conf,SubPath:dispersion.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-625x5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-ring-rebalance-v7q9j_openstack(9fdfdc9d-aa13-4fdb-8e1f-c54801e47867): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:53:32 crc kubenswrapper[4770]: E1209 11:53:32.722703 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/swift-ring-rebalance-v7q9j" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" Dec 09 11:53:32 crc kubenswrapper[4770]: I1209 11:53:32.841494 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.189739 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb\") pod \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.190104 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config\") pod \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.190184 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb\") pod \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.190263 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph6f7\" (UniqueName: \"kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7\") pod \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.190304 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc\") pod \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\" (UID: \"277b0a2e-73f7-4df2-9e01-d0145f26c13c\") " Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.199477 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7" (OuterVolumeSpecName: "kube-api-access-ph6f7") pod "277b0a2e-73f7-4df2-9e01-d0145f26c13c" (UID: "277b0a2e-73f7-4df2-9e01-d0145f26c13c"). InnerVolumeSpecName "kube-api-access-ph6f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.240677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "277b0a2e-73f7-4df2-9e01-d0145f26c13c" (UID: "277b0a2e-73f7-4df2-9e01-d0145f26c13c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.247936 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "277b0a2e-73f7-4df2-9e01-d0145f26c13c" (UID: "277b0a2e-73f7-4df2-9e01-d0145f26c13c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.249537 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config" (OuterVolumeSpecName: "config") pod "277b0a2e-73f7-4df2-9e01-d0145f26c13c" (UID: "277b0a2e-73f7-4df2-9e01-d0145f26c13c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.272239 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "277b0a2e-73f7-4df2-9e01-d0145f26c13c" (UID: "277b0a2e-73f7-4df2-9e01-d0145f26c13c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.288869 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhnq4-config-8rqdb"] Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.292470 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph6f7\" (UniqueName: \"kubernetes.io/projected/277b0a2e-73f7-4df2-9e01-d0145f26c13c-kube-api-access-ph6f7\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.292494 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.292503 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.292512 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.292520 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/277b0a2e-73f7-4df2-9e01-d0145f26c13c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:33 crc kubenswrapper[4770]: W1209 11:53:33.292930 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode357acf1_47f2_4aee_9fe3_79e5377711ae.slice/crio-668cbaa7256cff61c956aa680cfd0788900c8d064ebb9aec629070e4292d0768 WatchSource:0}: Error finding container 668cbaa7256cff61c956aa680cfd0788900c8d064ebb9aec629070e4292d0768: Status 404 returned error can't find the container with id 668cbaa7256cff61c956aa680cfd0788900c8d064ebb9aec629070e4292d0768 Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.632327 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.632318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-6mk5d" event={"ID":"277b0a2e-73f7-4df2-9e01-d0145f26c13c","Type":"ContainerDied","Data":"49759d3f2b7ec95c516039c5c263a6f1b333f84550c14fb0c80ea59aab56c403"} Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.632780 4770 scope.go:117] "RemoveContainer" containerID="64058d2e13281efc9c804ad49826b65727e1a62f615bf82bc8ad24bb102eaa4d" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.636599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerStarted","Data":"f2d4dc50bf2b36a841745128ed7fc9d07db8df1ab56a2af6bc7a83543d7f633f"} Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.636767 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.647780 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerStarted","Data":"1980ac00c990360135915c858d0f125e136b46aa153c5a4d5cf4019dd84f4355"} Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.648018 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.651688 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4-config-8rqdb" event={"ID":"e357acf1-47f2-4aee-9fe3-79e5377711ae","Type":"ContainerStarted","Data":"cd1933388921d9c04dede556a2c178263df5ffaf0f51165517bcc4205f9bbbac"} Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.651733 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4-config-8rqdb" event={"ID":"e357acf1-47f2-4aee-9fe3-79e5377711ae","Type":"ContainerStarted","Data":"668cbaa7256cff61c956aa680cfd0788900c8d064ebb9aec629070e4292d0768"} Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.656598 4770 scope.go:117] "RemoveContainer" containerID="1674d9400cde6a620f36617354b4773df843ea2997aafea12c9d03a72568c0b7" Dec 09 11:53:33 crc kubenswrapper[4770]: E1209 11:53:33.656751 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:ac7fefe1c93839c7ccb2aaa0a18751df0e9f64a36a3b4cc1b81d82d7774b8b45\\\"\"" pod="openstack/swift-ring-rebalance-v7q9j" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.696268 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=43.107307426 podStartE2EDuration="1m43.696248097s" podCreationTimestamp="2025-12-09 11:51:50 +0000 UTC" firstStartedPulling="2025-12-09 11:51:54.22864742 +0000 UTC m=+1239.469405939" lastFinishedPulling="2025-12-09 11:52:54.817588091 +0000 UTC m=+1300.058346610" observedRunningTime="2025-12-09 11:53:33.678312362 +0000 UTC m=+1338.919070891" watchObservedRunningTime="2025-12-09 11:53:33.696248097 +0000 UTC m=+1338.937006616" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.697043 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.703776 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-6mk5d"] Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.729310 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.294404129 podStartE2EDuration="1m44.729289954s" podCreationTimestamp="2025-12-09 11:51:49 +0000 UTC" firstStartedPulling="2025-12-09 11:51:54.335990282 +0000 UTC m=+1239.576748801" lastFinishedPulling="2025-12-09 11:52:54.770876107 +0000 UTC m=+1300.011634626" observedRunningTime="2025-12-09 11:53:33.720388209 +0000 UTC m=+1338.961146728" watchObservedRunningTime="2025-12-09 11:53:33.729289954 +0000 UTC m=+1338.970048473" Dec 09 11:53:33 crc kubenswrapper[4770]: I1209 11:53:33.742043 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xhnq4-config-8rqdb" podStartSLOduration=3.742021147 podStartE2EDuration="3.742021147s" podCreationTimestamp="2025-12-09 11:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:33.737282197 +0000 UTC m=+1338.978040726" watchObservedRunningTime="2025-12-09 11:53:33.742021147 +0000 UTC m=+1338.982779666" Dec 09 11:53:34 crc kubenswrapper[4770]: I1209 11:53:34.662030 4770 generic.go:334] "Generic (PLEG): container finished" podID="e357acf1-47f2-4aee-9fe3-79e5377711ae" containerID="cd1933388921d9c04dede556a2c178263df5ffaf0f51165517bcc4205f9bbbac" exitCode=0 Dec 09 11:53:34 crc kubenswrapper[4770]: I1209 11:53:34.662193 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4-config-8rqdb" event={"ID":"e357acf1-47f2-4aee-9fe3-79e5377711ae","Type":"ContainerDied","Data":"cd1933388921d9c04dede556a2c178263df5ffaf0f51165517bcc4205f9bbbac"} Dec 09 11:53:35 crc kubenswrapper[4770]: I1209 11:53:35.024449 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xhnq4" Dec 09 11:53:35 crc kubenswrapper[4770]: I1209 11:53:35.350085 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" path="/var/lib/kubelet/pods/277b0a2e-73f7-4df2-9e01-d0145f26c13c/volumes" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.021627 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.151882 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.151985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152396 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152446 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29bjp\" (UniqueName: \"kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152465 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152599 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn\") pod \"e357acf1-47f2-4aee-9fe3-79e5377711ae\" (UID: \"e357acf1-47f2-4aee-9fe3-79e5377711ae\") " Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152688 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run" (OuterVolumeSpecName: "var-run") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.152965 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153412 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153438 4770 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153456 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153495 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e357acf1-47f2-4aee-9fe3-79e5377711ae-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.153563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts" (OuterVolumeSpecName: "scripts") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.160163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp" (OuterVolumeSpecName: "kube-api-access-29bjp") pod "e357acf1-47f2-4aee-9fe3-79e5377711ae" (UID: "e357acf1-47f2-4aee-9fe3-79e5377711ae"). InnerVolumeSpecName "kube-api-access-29bjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.255019 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29bjp\" (UniqueName: \"kubernetes.io/projected/e357acf1-47f2-4aee-9fe3-79e5377711ae-kube-api-access-29bjp\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.255054 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e357acf1-47f2-4aee-9fe3-79e5377711ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.384571 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xhnq4-config-8rqdb"] Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.399217 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xhnq4-config-8rqdb"] Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.681106 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="668cbaa7256cff61c956aa680cfd0788900c8d064ebb9aec629070e4292d0768" Dec 09 11:53:36 crc kubenswrapper[4770]: I1209 11:53:36.681169 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4-config-8rqdb" Dec 09 11:53:37 crc kubenswrapper[4770]: I1209 11:53:37.352097 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e357acf1-47f2-4aee-9fe3-79e5377711ae" path="/var/lib/kubelet/pods/e357acf1-47f2-4aee-9fe3-79e5377711ae/volumes" Dec 09 11:53:41 crc kubenswrapper[4770]: I1209 11:53:41.756950 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:53:41 crc kubenswrapper[4770]: E1209 11:53:41.757268 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 11:53:41 crc kubenswrapper[4770]: E1209 11:53:41.757620 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 11:53:41 crc kubenswrapper[4770]: E1209 11:53:41.757683 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift podName:a0b34aa9-ae84-4b79-ba88-25c86cf5c51f nodeName:}" failed. No retries permitted until 2025-12-09 11:54:13.757663258 +0000 UTC m=+1378.998421777 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift") pod "swift-storage-0" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f") : configmap "swift-ring-files" not found Dec 09 11:53:42 crc kubenswrapper[4770]: I1209 11:53:42.515086 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Dec 09 11:53:42 crc kubenswrapper[4770]: I1209 11:53:42.550573 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Dec 09 11:53:45 crc kubenswrapper[4770]: I1209 11:53:45.753406 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-v7q9j" event={"ID":"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867","Type":"ContainerStarted","Data":"b70ca7d01822bd66dce685f79e1d82c3e5fbe093736ea216ac67d51a955e093b"} Dec 09 11:53:45 crc kubenswrapper[4770]: I1209 11:53:45.781223 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-v7q9j" podStartSLOduration=2.569243925 podStartE2EDuration="32.781204911s" podCreationTimestamp="2025-12-09 11:53:13 +0000 UTC" firstStartedPulling="2025-12-09 11:53:14.660801602 +0000 UTC m=+1319.901560121" lastFinishedPulling="2025-12-09 11:53:44.872762588 +0000 UTC m=+1350.113521107" observedRunningTime="2025-12-09 11:53:45.773934507 +0000 UTC m=+1351.014693026" watchObservedRunningTime="2025-12-09 11:53:45.781204911 +0000 UTC m=+1351.021963430" Dec 09 11:53:46 crc kubenswrapper[4770]: I1209 11:53:46.762771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pxdc5" event={"ID":"0c091c39-7a67-414c-9cbd-d323073c87be","Type":"ContainerStarted","Data":"bd7eefa592658dc5f5efab697a12a1e89949a1ef3faff590424391e7c9cbec12"} Dec 09 11:53:46 crc kubenswrapper[4770]: I1209 11:53:46.777399 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-pxdc5" podStartSLOduration=2.946534892 podStartE2EDuration="35.777380729s" podCreationTimestamp="2025-12-09 11:53:11 +0000 UTC" firstStartedPulling="2025-12-09 11:53:12.95217207 +0000 UTC m=+1318.192930599" lastFinishedPulling="2025-12-09 11:53:45.783017917 +0000 UTC m=+1351.023776436" observedRunningTime="2025-12-09 11:53:46.776316932 +0000 UTC m=+1352.017075471" watchObservedRunningTime="2025-12-09 11:53:46.777380729 +0000 UTC m=+1352.018139248" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.515361 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.551125 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.812454 4770 generic.go:334] "Generic (PLEG): container finished" podID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" containerID="b70ca7d01822bd66dce685f79e1d82c3e5fbe093736ea216ac67d51a955e093b" exitCode=0 Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.812514 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-v7q9j" event={"ID":"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867","Type":"ContainerDied","Data":"b70ca7d01822bd66dce685f79e1d82c3e5fbe093736ea216ac67d51a955e093b"} Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.839305 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-j4q62"] Dec 09 11:53:52 crc kubenswrapper[4770]: E1209 11:53:52.839695 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e357acf1-47f2-4aee-9fe3-79e5377711ae" containerName="ovn-config" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.839720 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e357acf1-47f2-4aee-9fe3-79e5377711ae" containerName="ovn-config" Dec 09 11:53:52 crc kubenswrapper[4770]: E1209 11:53:52.839758 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="init" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.839767 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="init" Dec 09 11:53:52 crc kubenswrapper[4770]: E1209 11:53:52.839779 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.839786 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.840036 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e357acf1-47f2-4aee-9fe3-79e5377711ae" containerName="ovn-config" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.840055 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b0a2e-73f7-4df2-9e01-d0145f26c13c" containerName="dnsmasq-dns" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.840729 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.893447 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-j4q62"] Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.922604 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p57bd\" (UniqueName: \"kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:52 crc kubenswrapper[4770]: I1209 11:53:52.922723 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.025680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.025887 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p57bd\" (UniqueName: \"kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.026554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.054310 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lrtgq"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.055583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.092439 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p57bd\" (UniqueName: \"kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd\") pod \"barbican-db-create-j4q62\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.114796 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-aaaf-account-create-update-bv8tm"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.116427 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.119144 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.193060 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.225159 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aaaf-account-create-update-bv8tm"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.228558 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhrmn\" (UniqueName: \"kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.228677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.228729 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.228749 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tlgd\" (UniqueName: \"kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.267005 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lrtgq"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.733556 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.733630 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tlgd\" (UniqueName: \"kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.733985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhrmn\" (UniqueName: \"kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.751753 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.754794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.756065 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.786259 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tlgd\" (UniqueName: \"kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd\") pod \"cinder-db-create-lrtgq\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.795452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhrmn\" (UniqueName: \"kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn\") pod \"barbican-aaaf-account-create-update-bv8tm\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.819717 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.893199 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4e4b-account-create-update-shln7"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.896369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.908051 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4e4b-account-create-update-shln7"] Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.909319 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.959094 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fc6j\" (UniqueName: \"kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.959183 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:53 crc kubenswrapper[4770]: I1209 11:53:53.978366 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.009278 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jh8bg"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.010487 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.026009 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sz4h5" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.028696 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.029026 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.035794 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.044800 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-z28k6"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.048792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.061356 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.061405 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fc6j\" (UniqueName: \"kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.061497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.061525 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvcn\" (UniqueName: \"kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.061557 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.063004 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.102780 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jh8bg"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.141696 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fc6j\" (UniqueName: \"kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j\") pod \"cinder-4e4b-account-create-update-shln7\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.149292 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z28k6"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.163244 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.163324 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.163392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.163418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcr6v\" (UniqueName: \"kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.163445 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvcn\" (UniqueName: \"kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.178922 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.179615 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.205138 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvcn\" (UniqueName: \"kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn\") pod \"keystone-db-sync-jh8bg\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.265272 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcr6v\" (UniqueName: \"kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.265493 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.266605 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.290015 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.307018 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2ce4-account-create-update-7d6qp"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.308441 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.321838 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.326647 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcr6v\" (UniqueName: \"kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v\") pod \"neutron-db-create-z28k6\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.337526 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.356108 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2ce4-account-create-update-7d6qp"] Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.368066 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqn8d\" (UniqueName: \"kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.368255 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.378706 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.471144 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqn8d\" (UniqueName: \"kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.471592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.482483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.558604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqn8d\" (UniqueName: \"kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d\") pod \"neutron-2ce4-account-create-update-7d6qp\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.794399 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.854212 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.857194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-v7q9j" event={"ID":"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867","Type":"ContainerDied","Data":"eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3"} Dec 09 11:53:54 crc kubenswrapper[4770]: I1209 11:53:54.857236 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eba3216ebb3a5b7829d06a59ec3fee84b361287a63436a813708aaef3c1962e3" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.982921 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-j4q62"] Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.992855 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.992947 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.993012 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.993097 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.993137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.993224 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:54.993434 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-625x5\" (UniqueName: \"kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5\") pod \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\" (UID: \"9fdfdc9d-aa13-4fdb-8e1f-c54801e47867\") " Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.298030 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.298409 4770 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.306738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.324461 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5" (OuterVolumeSpecName: "kube-api-access-625x5") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "kube-api-access-625x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.404561 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-625x5\" (UniqueName: \"kubernetes.io/projected/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-kube-api-access-625x5\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.404600 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.406259 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.453057 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.498094 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aaaf-account-create-update-bv8tm"] Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.508042 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts" (OuterVolumeSpecName: "scripts") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.511021 4770 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.511052 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.511063 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.523709 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" (UID: "9fdfdc9d-aa13-4fdb-8e1f-c54801e47867"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.607148 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4e4b-account-create-update-shln7"] Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.612988 4770 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.726665 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lrtgq"] Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.797732 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jh8bg"] Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.825345 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z28k6"] Dec 09 11:53:55 crc kubenswrapper[4770]: W1209 11:53:55.847636 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5c87b1_e0b3_4455_a2ad_a14e5bd241ae.slice/crio-a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5 WatchSource:0}: Error finding container a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5: Status 404 returned error can't find the container with id a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5 Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.878771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaaf-account-create-update-bv8tm" event={"ID":"944b3781-77eb-4993-a42d-f483bb90ffea","Type":"ContainerStarted","Data":"9941dbcdc944c557b193fbf074adcf6381fa10a6d213fce09f09f89ac8040b58"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.878819 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaaf-account-create-update-bv8tm" event={"ID":"944b3781-77eb-4993-a42d-f483bb90ffea","Type":"ContainerStarted","Data":"009ef72ccf5f4f24a1d0aae8dd041d97db684a896f9ccfbed49d5dbedfabb91e"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.887026 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z28k6" event={"ID":"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae","Type":"ContainerStarted","Data":"a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.889025 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lrtgq" event={"ID":"0fc164b2-f1a7-4c38-be49-f53427b7fd91","Type":"ContainerStarted","Data":"fe42639b2c6c624b16219dc2e71427d56a0f13e7569f1030ed8d6df62a599d68"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.901652 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-j4q62" event={"ID":"00b4541a-003a-4282-ad4e-4b1f24106f15","Type":"ContainerStarted","Data":"c22d19cca7b2713787bd65a3a2f28d2818e0752a37bcb6ec50e1587fb6477014"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.901711 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-j4q62" event={"ID":"00b4541a-003a-4282-ad4e-4b1f24106f15","Type":"ContainerStarted","Data":"6a3a97979711c9b22c98e79f34e80166d674c261409dd34f3fb3c49277b96425"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.910244 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jh8bg" event={"ID":"109b0cbe-45dc-416c-a4a2-972eda68baf2","Type":"ContainerStarted","Data":"fb82397051e2efe8937101c7f0d87d4bec628a681dca4a40f9c2babd8152a96f"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.915914 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-aaaf-account-create-update-bv8tm" podStartSLOduration=2.915877008 podStartE2EDuration="2.915877008s" podCreationTimestamp="2025-12-09 11:53:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:55.900832447 +0000 UTC m=+1361.141590966" watchObservedRunningTime="2025-12-09 11:53:55.915877008 +0000 UTC m=+1361.156635527" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.919801 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-v7q9j" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.919945 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4e4b-account-create-update-shln7" event={"ID":"aad01631-8fd1-42c6-abcd-989631e88fd5","Type":"ContainerStarted","Data":"7d1eb31fddc9b44fd9d13bf2118bc46c56e247d2368c5da71f2f49e4c1c3bb1c"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.919989 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4e4b-account-create-update-shln7" event={"ID":"aad01631-8fd1-42c6-abcd-989631e88fd5","Type":"ContainerStarted","Data":"062cb279a94e3c638ea408ab3438f45c38915366267d2939da2a818575b6072f"} Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.940251 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-j4q62" podStartSLOduration=3.9402114150000003 podStartE2EDuration="3.940211415s" podCreationTimestamp="2025-12-09 11:53:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:55.929248497 +0000 UTC m=+1361.170007016" watchObservedRunningTime="2025-12-09 11:53:55.940211415 +0000 UTC m=+1361.180969944" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.979165 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-4e4b-account-create-update-shln7" podStartSLOduration=2.979142742 podStartE2EDuration="2.979142742s" podCreationTimestamp="2025-12-09 11:53:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:53:55.966888991 +0000 UTC m=+1361.207647510" watchObservedRunningTime="2025-12-09 11:53:55.979142742 +0000 UTC m=+1361.219901261" Dec 09 11:53:55 crc kubenswrapper[4770]: I1209 11:53:55.979986 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2ce4-account-create-update-7d6qp"] Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.929941 4770 generic.go:334] "Generic (PLEG): container finished" podID="d663e7bf-6267-4d24-a23b-ba2e6924bd0e" containerID="b090a5d51438df67dcd20b0f78e7f2772fbbf4a6a66ab96fd227723a8cab51f7" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.930392 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2ce4-account-create-update-7d6qp" event={"ID":"d663e7bf-6267-4d24-a23b-ba2e6924bd0e","Type":"ContainerDied","Data":"b090a5d51438df67dcd20b0f78e7f2772fbbf4a6a66ab96fd227723a8cab51f7"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.930429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2ce4-account-create-update-7d6qp" event={"ID":"d663e7bf-6267-4d24-a23b-ba2e6924bd0e","Type":"ContainerStarted","Data":"440313c37ee599ef0d9620cb8d05358ce1a16e671fb076face8c768c04807d4d"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.932374 4770 generic.go:334] "Generic (PLEG): container finished" podID="944b3781-77eb-4993-a42d-f483bb90ffea" containerID="9941dbcdc944c557b193fbf074adcf6381fa10a6d213fce09f09f89ac8040b58" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.932433 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaaf-account-create-update-bv8tm" event={"ID":"944b3781-77eb-4993-a42d-f483bb90ffea","Type":"ContainerDied","Data":"9941dbcdc944c557b193fbf074adcf6381fa10a6d213fce09f09f89ac8040b58"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.934925 4770 generic.go:334] "Generic (PLEG): container finished" podID="0c091c39-7a67-414c-9cbd-d323073c87be" containerID="bd7eefa592658dc5f5efab697a12a1e89949a1ef3faff590424391e7c9cbec12" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.935016 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pxdc5" event={"ID":"0c091c39-7a67-414c-9cbd-d323073c87be","Type":"ContainerDied","Data":"bd7eefa592658dc5f5efab697a12a1e89949a1ef3faff590424391e7c9cbec12"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.936748 4770 generic.go:334] "Generic (PLEG): container finished" podID="9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" containerID="84ccb36f1c40cfd48e74cef605f7a217a151d292ac3cb99fbc76392a343e5ec9" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.936805 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z28k6" event={"ID":"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae","Type":"ContainerDied","Data":"84ccb36f1c40cfd48e74cef605f7a217a151d292ac3cb99fbc76392a343e5ec9"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.938623 4770 generic.go:334] "Generic (PLEG): container finished" podID="0fc164b2-f1a7-4c38-be49-f53427b7fd91" containerID="4123eed8acdc05741e19a06c69e7357fb168f03ee4b24e968baf996a6cf73961" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.938719 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lrtgq" event={"ID":"0fc164b2-f1a7-4c38-be49-f53427b7fd91","Type":"ContainerDied","Data":"4123eed8acdc05741e19a06c69e7357fb168f03ee4b24e968baf996a6cf73961"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.940679 4770 generic.go:334] "Generic (PLEG): container finished" podID="00b4541a-003a-4282-ad4e-4b1f24106f15" containerID="c22d19cca7b2713787bd65a3a2f28d2818e0752a37bcb6ec50e1587fb6477014" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.940758 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-j4q62" event={"ID":"00b4541a-003a-4282-ad4e-4b1f24106f15","Type":"ContainerDied","Data":"c22d19cca7b2713787bd65a3a2f28d2818e0752a37bcb6ec50e1587fb6477014"} Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.945062 4770 generic.go:334] "Generic (PLEG): container finished" podID="aad01631-8fd1-42c6-abcd-989631e88fd5" containerID="7d1eb31fddc9b44fd9d13bf2118bc46c56e247d2368c5da71f2f49e4c1c3bb1c" exitCode=0 Dec 09 11:53:56 crc kubenswrapper[4770]: I1209 11:53:56.945130 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4e4b-account-create-update-shln7" event={"ID":"aad01631-8fd1-42c6-abcd-989631e88fd5","Type":"ContainerDied","Data":"7d1eb31fddc9b44fd9d13bf2118bc46c56e247d2368c5da71f2f49e4c1c3bb1c"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.446851 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.556245 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts\") pod \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.556386 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcr6v\" (UniqueName: \"kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v\") pod \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\" (UID: \"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.556798 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" (UID: "9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.566548 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v" (OuterVolumeSpecName: "kube-api-access-wcr6v") pod "9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" (UID: "9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae"). InnerVolumeSpecName "kube-api-access-wcr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.640223 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.653272 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.657791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts\") pod \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.657974 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tlgd\" (UniqueName: \"kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd\") pod \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\" (UID: \"0fc164b2-f1a7-4c38-be49-f53427b7fd91\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.658783 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0fc164b2-f1a7-4c38-be49-f53427b7fd91" (UID: "0fc164b2-f1a7-4c38-be49-f53427b7fd91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.659267 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fc164b2-f1a7-4c38-be49-f53427b7fd91-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.659285 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.659297 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcr6v\" (UniqueName: \"kubernetes.io/projected/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae-kube-api-access-wcr6v\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.659764 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.661854 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd" (OuterVolumeSpecName: "kube-api-access-9tlgd") pod "0fc164b2-f1a7-4c38-be49-f53427b7fd91" (UID: "0fc164b2-f1a7-4c38-be49-f53427b7fd91"). InnerVolumeSpecName "kube-api-access-9tlgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.667359 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.715517 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760108 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts\") pod \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760155 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fc6j\" (UniqueName: \"kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j\") pod \"aad01631-8fd1-42c6-abcd-989631e88fd5\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760236 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts\") pod \"aad01631-8fd1-42c6-abcd-989631e88fd5\" (UID: \"aad01631-8fd1-42c6-abcd-989631e88fd5\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts\") pod \"00b4541a-003a-4282-ad4e-4b1f24106f15\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760346 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqn8d\" (UniqueName: \"kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d\") pod \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\" (UID: \"d663e7bf-6267-4d24-a23b-ba2e6924bd0e\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760413 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p57bd\" (UniqueName: \"kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd\") pod \"00b4541a-003a-4282-ad4e-4b1f24106f15\" (UID: \"00b4541a-003a-4282-ad4e-4b1f24106f15\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760436 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts\") pod \"944b3781-77eb-4993-a42d-f483bb90ffea\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760473 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhrmn\" (UniqueName: \"kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn\") pod \"944b3781-77eb-4993-a42d-f483bb90ffea\" (UID: \"944b3781-77eb-4993-a42d-f483bb90ffea\") " Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760937 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d663e7bf-6267-4d24-a23b-ba2e6924bd0e" (UID: "d663e7bf-6267-4d24-a23b-ba2e6924bd0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.760990 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tlgd\" (UniqueName: \"kubernetes.io/projected/0fc164b2-f1a7-4c38-be49-f53427b7fd91-kube-api-access-9tlgd\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.761314 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00b4541a-003a-4282-ad4e-4b1f24106f15" (UID: "00b4541a-003a-4282-ad4e-4b1f24106f15"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.761387 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "944b3781-77eb-4993-a42d-f483bb90ffea" (UID: "944b3781-77eb-4993-a42d-f483bb90ffea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.761446 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aad01631-8fd1-42c6-abcd-989631e88fd5" (UID: "aad01631-8fd1-42c6-abcd-989631e88fd5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.764942 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d" (OuterVolumeSpecName: "kube-api-access-jqn8d") pod "d663e7bf-6267-4d24-a23b-ba2e6924bd0e" (UID: "d663e7bf-6267-4d24-a23b-ba2e6924bd0e"). InnerVolumeSpecName "kube-api-access-jqn8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.764991 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j" (OuterVolumeSpecName: "kube-api-access-4fc6j") pod "aad01631-8fd1-42c6-abcd-989631e88fd5" (UID: "aad01631-8fd1-42c6-abcd-989631e88fd5"). InnerVolumeSpecName "kube-api-access-4fc6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.766680 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn" (OuterVolumeSpecName: "kube-api-access-vhrmn") pod "944b3781-77eb-4993-a42d-f483bb90ffea" (UID: "944b3781-77eb-4993-a42d-f483bb90ffea"). InnerVolumeSpecName "kube-api-access-vhrmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.770911 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd" (OuterVolumeSpecName: "kube-api-access-p57bd") pod "00b4541a-003a-4282-ad4e-4b1f24106f15" (UID: "00b4541a-003a-4282-ad4e-4b1f24106f15"). InnerVolumeSpecName "kube-api-access-p57bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863226 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aad01631-8fd1-42c6-abcd-989631e88fd5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863264 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00b4541a-003a-4282-ad4e-4b1f24106f15-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863276 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqn8d\" (UniqueName: \"kubernetes.io/projected/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-kube-api-access-jqn8d\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863285 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p57bd\" (UniqueName: \"kubernetes.io/projected/00b4541a-003a-4282-ad4e-4b1f24106f15-kube-api-access-p57bd\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863294 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944b3781-77eb-4993-a42d-f483bb90ffea-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863304 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhrmn\" (UniqueName: \"kubernetes.io/projected/944b3781-77eb-4993-a42d-f483bb90ffea-kube-api-access-vhrmn\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863314 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d663e7bf-6267-4d24-a23b-ba2e6924bd0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.863324 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fc6j\" (UniqueName: \"kubernetes.io/projected/aad01631-8fd1-42c6-abcd-989631e88fd5-kube-api-access-4fc6j\") on node \"crc\" DevicePath \"\"" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.966681 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaaf-account-create-update-bv8tm" event={"ID":"944b3781-77eb-4993-a42d-f483bb90ffea","Type":"ContainerDied","Data":"009ef72ccf5f4f24a1d0aae8dd041d97db684a896f9ccfbed49d5dbedfabb91e"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.966696 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaaf-account-create-update-bv8tm" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.966716 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="009ef72ccf5f4f24a1d0aae8dd041d97db684a896f9ccfbed49d5dbedfabb91e" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.969155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z28k6" event={"ID":"9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae","Type":"ContainerDied","Data":"a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.969175 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z28k6" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.969197 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73b8584282756a1095b206612eff36d124b90f71379442e8507d3988a4e40e5" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.970933 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lrtgq" event={"ID":"0fc164b2-f1a7-4c38-be49-f53427b7fd91","Type":"ContainerDied","Data":"fe42639b2c6c624b16219dc2e71427d56a0f13e7569f1030ed8d6df62a599d68"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.970961 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe42639b2c6c624b16219dc2e71427d56a0f13e7569f1030ed8d6df62a599d68" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.970983 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lrtgq" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.972303 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-j4q62" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.972322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-j4q62" event={"ID":"00b4541a-003a-4282-ad4e-4b1f24106f15","Type":"ContainerDied","Data":"6a3a97979711c9b22c98e79f34e80166d674c261409dd34f3fb3c49277b96425"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.972350 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3a97979711c9b22c98e79f34e80166d674c261409dd34f3fb3c49277b96425" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.977804 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4e4b-account-create-update-shln7" event={"ID":"aad01631-8fd1-42c6-abcd-989631e88fd5","Type":"ContainerDied","Data":"062cb279a94e3c638ea408ab3438f45c38915366267d2939da2a818575b6072f"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.977825 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="062cb279a94e3c638ea408ab3438f45c38915366267d2939da2a818575b6072f" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.977866 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4e4b-account-create-update-shln7" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.983109 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2ce4-account-create-update-7d6qp" event={"ID":"d663e7bf-6267-4d24-a23b-ba2e6924bd0e","Type":"ContainerDied","Data":"440313c37ee599ef0d9620cb8d05358ce1a16e671fb076face8c768c04807d4d"} Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.983136 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="440313c37ee599ef0d9620cb8d05358ce1a16e671fb076face8c768c04807d4d" Dec 09 11:53:59 crc kubenswrapper[4770]: I1209 11:53:58.983178 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2ce4-account-create-update-7d6qp" Dec 09 11:54:02 crc kubenswrapper[4770]: I1209 11:54:02.475985 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:54:02 crc kubenswrapper[4770]: I1209 11:54:02.477304 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:54:02 crc kubenswrapper[4770]: I1209 11:54:02.994502 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pxdc5" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.090475 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whvds\" (UniqueName: \"kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds\") pod \"0c091c39-7a67-414c-9cbd-d323073c87be\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.090934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data\") pod \"0c091c39-7a67-414c-9cbd-d323073c87be\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.091061 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data\") pod \"0c091c39-7a67-414c-9cbd-d323073c87be\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.091162 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle\") pod \"0c091c39-7a67-414c-9cbd-d323073c87be\" (UID: \"0c091c39-7a67-414c-9cbd-d323073c87be\") " Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.096649 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds" (OuterVolumeSpecName: "kube-api-access-whvds") pod "0c091c39-7a67-414c-9cbd-d323073c87be" (UID: "0c091c39-7a67-414c-9cbd-d323073c87be"). InnerVolumeSpecName "kube-api-access-whvds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.097770 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0c091c39-7a67-414c-9cbd-d323073c87be" (UID: "0c091c39-7a67-414c-9cbd-d323073c87be"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.102424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pxdc5" event={"ID":"0c091c39-7a67-414c-9cbd-d323073c87be","Type":"ContainerDied","Data":"c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f"} Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.102468 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7f4d058050a1c5fd294eaba26a9ad16d3886717a73afe6f1cd24c58485a034f" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.102522 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pxdc5" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.119986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c091c39-7a67-414c-9cbd-d323073c87be" (UID: "0c091c39-7a67-414c-9cbd-d323073c87be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.146097 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data" (OuterVolumeSpecName: "config-data") pod "0c091c39-7a67-414c-9cbd-d323073c87be" (UID: "0c091c39-7a67-414c-9cbd-d323073c87be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.192533 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.192585 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.192599 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c091c39-7a67-414c-9cbd-d323073c87be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:03 crc kubenswrapper[4770]: I1209 11:54:03.192613 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whvds\" (UniqueName: \"kubernetes.io/projected/0c091c39-7a67-414c-9cbd-d323073c87be-kube-api-access-whvds\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.118894 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jh8bg" event={"ID":"109b0cbe-45dc-416c-a4a2-972eda68baf2","Type":"ContainerStarted","Data":"036cd9d85dac93e7bd1195800ae605a32e7659f3dc8866a12f69f81295bfb20c"} Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.137459 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jh8bg" podStartSLOduration=3.8824992959999998 podStartE2EDuration="11.137436868s" podCreationTimestamp="2025-12-09 11:53:53 +0000 UTC" firstStartedPulling="2025-12-09 11:53:55.834993957 +0000 UTC m=+1361.075752476" lastFinishedPulling="2025-12-09 11:54:03.089931529 +0000 UTC m=+1368.330690048" observedRunningTime="2025-12-09 11:54:04.134805241 +0000 UTC m=+1369.375563760" watchObservedRunningTime="2025-12-09 11:54:04.137436868 +0000 UTC m=+1369.378195387" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.456881 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457354 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944b3781-77eb-4993-a42d-f483bb90ffea" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457380 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="944b3781-77eb-4993-a42d-f483bb90ffea" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457402 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" containerName="glance-db-sync" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457411 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" containerName="glance-db-sync" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457422 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d663e7bf-6267-4d24-a23b-ba2e6924bd0e" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457431 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d663e7bf-6267-4d24-a23b-ba2e6924bd0e" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457443 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" containerName="swift-ring-rebalance" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457451 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" containerName="swift-ring-rebalance" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457467 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00b4541a-003a-4282-ad4e-4b1f24106f15" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457473 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="00b4541a-003a-4282-ad4e-4b1f24106f15" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457493 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad01631-8fd1-42c6-abcd-989631e88fd5" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457501 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad01631-8fd1-42c6-abcd-989631e88fd5" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457516 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457523 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: E1209 11:54:04.457536 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fc164b2-f1a7-4c38-be49-f53427b7fd91" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457543 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fc164b2-f1a7-4c38-be49-f53427b7fd91" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457748 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fc164b2-f1a7-4c38-be49-f53427b7fd91" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457795 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" containerName="swift-ring-rebalance" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457813 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="944b3781-77eb-4993-a42d-f483bb90ffea" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457838 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457847 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d663e7bf-6267-4d24-a23b-ba2e6924bd0e" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457870 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad01631-8fd1-42c6-abcd-989631e88fd5" containerName="mariadb-account-create-update" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457882 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" containerName="glance-db-sync" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.457930 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="00b4541a-003a-4282-ad4e-4b1f24106f15" containerName="mariadb-database-create" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.459167 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.480943 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.675271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fc9\" (UniqueName: \"kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.675499 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.675582 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.675769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.675841 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.777040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fc9\" (UniqueName: \"kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.777183 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778088 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778145 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778270 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778324 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.778872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.779599 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:04 crc kubenswrapper[4770]: I1209 11:54:04.802333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fc9\" (UniqueName: \"kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9\") pod \"dnsmasq-dns-6bfd654465-rlbn9\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:05 crc kubenswrapper[4770]: I1209 11:54:05.092984 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:05 crc kubenswrapper[4770]: I1209 11:54:05.575097 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:06 crc kubenswrapper[4770]: I1209 11:54:06.151716 4770 generic.go:334] "Generic (PLEG): container finished" podID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerID="3b1e69bfca81ba9818f89c2292d3122c0b8f69fb6d133975c886d4ba60910931" exitCode=0 Dec 09 11:54:06 crc kubenswrapper[4770]: I1209 11:54:06.151829 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" event={"ID":"f6b8c6ab-f03c-4927-8a49-c174a34e1787","Type":"ContainerDied","Data":"3b1e69bfca81ba9818f89c2292d3122c0b8f69fb6d133975c886d4ba60910931"} Dec 09 11:54:06 crc kubenswrapper[4770]: I1209 11:54:06.152010 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" event={"ID":"f6b8c6ab-f03c-4927-8a49-c174a34e1787","Type":"ContainerStarted","Data":"8a7048f8b7fc46e7e0c35862b4ecc89e22e16220ec0e120deb34477c072f7fce"} Dec 09 11:54:07 crc kubenswrapper[4770]: I1209 11:54:07.171068 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" event={"ID":"f6b8c6ab-f03c-4927-8a49-c174a34e1787","Type":"ContainerStarted","Data":"24a5bfa21cd137763d9856bd661f767cc50715aade649f45bd52d1da769ae38a"} Dec 09 11:54:07 crc kubenswrapper[4770]: I1209 11:54:07.171457 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:07 crc kubenswrapper[4770]: I1209 11:54:07.200466 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" podStartSLOduration=3.200445304 podStartE2EDuration="3.200445304s" podCreationTimestamp="2025-12-09 11:54:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:07.190105231 +0000 UTC m=+1372.430863750" watchObservedRunningTime="2025-12-09 11:54:07.200445304 +0000 UTC m=+1372.441203823" Dec 09 11:54:10 crc kubenswrapper[4770]: I1209 11:54:10.198720 4770 generic.go:334] "Generic (PLEG): container finished" podID="109b0cbe-45dc-416c-a4a2-972eda68baf2" containerID="036cd9d85dac93e7bd1195800ae605a32e7659f3dc8866a12f69f81295bfb20c" exitCode=0 Dec 09 11:54:10 crc kubenswrapper[4770]: I1209 11:54:10.198834 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jh8bg" event={"ID":"109b0cbe-45dc-416c-a4a2-972eda68baf2","Type":"ContainerDied","Data":"036cd9d85dac93e7bd1195800ae605a32e7659f3dc8866a12f69f81295bfb20c"} Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.555153 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.754397 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbvcn\" (UniqueName: \"kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn\") pod \"109b0cbe-45dc-416c-a4a2-972eda68baf2\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.754461 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data\") pod \"109b0cbe-45dc-416c-a4a2-972eda68baf2\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.754513 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle\") pod \"109b0cbe-45dc-416c-a4a2-972eda68baf2\" (UID: \"109b0cbe-45dc-416c-a4a2-972eda68baf2\") " Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.761323 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn" (OuterVolumeSpecName: "kube-api-access-vbvcn") pod "109b0cbe-45dc-416c-a4a2-972eda68baf2" (UID: "109b0cbe-45dc-416c-a4a2-972eda68baf2"). InnerVolumeSpecName "kube-api-access-vbvcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.782114 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "109b0cbe-45dc-416c-a4a2-972eda68baf2" (UID: "109b0cbe-45dc-416c-a4a2-972eda68baf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.800783 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data" (OuterVolumeSpecName: "config-data") pod "109b0cbe-45dc-416c-a4a2-972eda68baf2" (UID: "109b0cbe-45dc-416c-a4a2-972eda68baf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.867211 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbvcn\" (UniqueName: \"kubernetes.io/projected/109b0cbe-45dc-416c-a4a2-972eda68baf2-kube-api-access-vbvcn\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.867473 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:11 crc kubenswrapper[4770]: I1209 11:54:11.867544 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109b0cbe-45dc-416c-a4a2-972eda68baf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.216576 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jh8bg" event={"ID":"109b0cbe-45dc-416c-a4a2-972eda68baf2","Type":"ContainerDied","Data":"fb82397051e2efe8937101c7f0d87d4bec628a681dca4a40f9c2babd8152a96f"} Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.216630 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb82397051e2efe8937101c7f0d87d4bec628a681dca4a40f9c2babd8152a96f" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.216697 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jh8bg" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.536524 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wr87b"] Dec 09 11:54:12 crc kubenswrapper[4770]: E1209 11:54:12.537330 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109b0cbe-45dc-416c-a4a2-972eda68baf2" containerName="keystone-db-sync" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.537354 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="109b0cbe-45dc-416c-a4a2-972eda68baf2" containerName="keystone-db-sync" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.537641 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="109b0cbe-45dc-416c-a4a2-972eda68baf2" containerName="keystone-db-sync" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.538366 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.543695 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.543790 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.543799 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.543729 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.543730 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sz4h5" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.614099 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wr87b"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.656962 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.690577 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.660967 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="dnsmasq-dns" containerID="cri-o://24a5bfa21cd137763d9856bd661f767cc50715aade649f45bd52d1da769ae38a" gracePeriod=10 Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.693657 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.693727 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.693803 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx924\" (UniqueName: \"kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.693857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.693885 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.694034 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.769529 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.771786 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.795847 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.795933 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.795986 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.796052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx924\" (UniqueName: \"kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.796097 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.796120 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.799325 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.812410 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.812460 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.812841 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.813490 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.819983 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.822636 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4m5vj"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.832938 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.844459 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx924\" (UniqueName: \"kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924\") pod \"keystone-bootstrap-wr87b\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.844864 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.845067 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.848340 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pxvc8" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.891992 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4m5vj"] Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.905337 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.906708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkxv7\" (UniqueName: \"kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.906757 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.906796 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.906861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:12 crc kubenswrapper[4770]: I1209 11:54:12.906950 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012630 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvr4\" (UniqueName: \"kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012875 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012928 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.012984 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkxv7\" (UniqueName: \"kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.013023 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.013048 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.013122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.013155 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.013198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.017121 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.017196 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-p2vmw"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.017483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.018964 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.020229 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.025708 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.028085 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.028390 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qkbpv" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.060350 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p2vmw"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.114771 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvr4\" (UniqueName: \"kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.114880 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.114935 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.114995 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.115081 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbwlz\" (UniqueName: \"kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.115174 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.115211 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.115256 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.115282 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.120162 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.120278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.130488 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkxv7\" (UniqueName: \"kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7\") pod \"dnsmasq-dns-99559fbf5-kfgdk\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.135011 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.138806 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.141823 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.142802 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.177333 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-l9l6m"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.178814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.185293 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6tnc9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.185533 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.185562 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.186048 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.204309 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvr4\" (UniqueName: \"kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4\") pod \"cinder-db-sync-4m5vj\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.211968 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-l9l6m"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.217615 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.217685 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.217852 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbwlz\" (UniqueName: \"kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.222229 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.222767 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.240437 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.275017 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.277059 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.280574 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbwlz\" (UniqueName: \"kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz\") pod \"barbican-db-sync-p2vmw\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.319678 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.319777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.319855 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hfpg\" (UniqueName: \"kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.328100 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zjjm9"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.330642 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.344357 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.344793 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.345142 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5tc9t" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.361239 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.366067 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zjjm9"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.376117 4770 generic.go:334] "Generic (PLEG): container finished" podID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerID="24a5bfa21cd137763d9856bd661f767cc50715aade649f45bd52d1da769ae38a" exitCode=0 Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.376164 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" event={"ID":"f6b8c6ab-f03c-4927-8a49-c174a34e1787","Type":"ContainerDied","Data":"24a5bfa21cd137763d9856bd661f767cc50715aade649f45bd52d1da769ae38a"} Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423425 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hfpg\" (UniqueName: \"kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423531 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qng\" (UniqueName: \"kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423586 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423612 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423642 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b26x4\" (UniqueName: \"kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423671 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423733 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423755 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423778 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423796 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.423833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.426344 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.430360 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.431446 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.435832 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.442804 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.451889 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.452395 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.479056 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hfpg\" (UniqueName: \"kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg\") pod \"neutron-db-sync-l9l6m\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.518068 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.527380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.527698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.527798 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.527922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528081 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528225 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528318 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528412 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qng\" (UniqueName: \"kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528670 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528711 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.528735 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.529355 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.529599 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533630 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533646 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533768 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533837 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc2sk\" (UniqueName: \"kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.533965 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.534153 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b26x4\" (UniqueName: \"kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.534322 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.535136 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.540224 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.546956 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.550689 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qng\" (UniqueName: \"kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng\") pod \"placement-db-sync-zjjm9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.561660 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b26x4\" (UniqueName: \"kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4\") pod \"dnsmasq-dns-67f84f7cd9-k9wdm\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.576291 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.623891 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.636415 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.637569 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.637670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.637751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.637814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc2sk\" (UniqueName: \"kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.637893 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.638131 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.639272 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.641239 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.645555 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.645766 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.646714 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.655653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.660949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc2sk\" (UniqueName: \"kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk\") pod \"ceilometer-0\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.665262 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.732696 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.739229 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.746847 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.747028 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fvqzp" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.747246 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.747582 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.791574 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wr87b"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.819538 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.841604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842616 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842676 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842698 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842728 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842776 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842849 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.842884 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.844614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.844691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjxw\" (UniqueName: \"kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.859823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"swift-storage-0\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " pod="openstack/swift-storage-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951467 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951515 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951549 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951572 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjxw\" (UniqueName: \"kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951692 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951709 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951726 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.951744 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.955174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.960390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.960655 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.960858 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.976274 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjxw\" (UniqueName: \"kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:13 crc kubenswrapper[4770]: I1209 11:54:13.978770 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.008036 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.008518 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.013113 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.015470 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.019835 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.020823 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.045673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.070360 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.084499 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.104121 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.120206 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.156112 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.156184 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.156235 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjvr6\" (UniqueName: \"kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157106 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157266 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157317 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157370 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.157857 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4m5vj"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.200488 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.261581 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config\") pod \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.261636 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb\") pod \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.261745 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb\") pod \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.261769 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2fc9\" (UniqueName: \"kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9\") pod \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.261825 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc\") pod \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\" (UID: \"f6b8c6ab-f03c-4927-8a49-c174a34e1787\") " Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262608 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjvr6\" (UniqueName: \"kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262653 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262782 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262854 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262929 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.262963 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.263624 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.267349 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.267409 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.275243 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9" (OuterVolumeSpecName: "kube-api-access-l2fc9") pod "f6b8c6ab-f03c-4927-8a49-c174a34e1787" (UID: "f6b8c6ab-f03c-4927-8a49-c174a34e1787"). InnerVolumeSpecName "kube-api-access-l2fc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.275520 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.276750 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.293006 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.296357 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.297428 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjvr6\" (UniqueName: \"kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.365939 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2fc9\" (UniqueName: \"kubernetes.io/projected/f6b8c6ab-f03c-4927-8a49-c174a34e1787-kube-api-access-l2fc9\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.452572 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" event={"ID":"f6b8c6ab-f03c-4927-8a49-c174a34e1787","Type":"ContainerDied","Data":"8a7048f8b7fc46e7e0c35862b4ecc89e22e16220ec0e120deb34477c072f7fce"} Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.452632 4770 scope.go:117] "RemoveContainer" containerID="24a5bfa21cd137763d9856bd661f767cc50715aade649f45bd52d1da769ae38a" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.453121 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bfd654465-rlbn9" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.505886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wr87b" event={"ID":"289530c3-8968-4421-91b9-325d95b941c4","Type":"ContainerStarted","Data":"ee1f899fe60cc1c9c5521e75450ed62da3beb63a31782c7b57617f9acd978c03"} Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.548048 4770 scope.go:117] "RemoveContainer" containerID="3b1e69bfca81ba9818f89c2292d3122c0b8f69fb6d133975c886d4ba60910931" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.548318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" event={"ID":"03c1adca-9233-4b79-a995-1d7e498d08a5","Type":"ContainerStarted","Data":"21121f0cec7f4ee54be03c3911e5b8ec6d290d5a8c4ce1817d287efb4a80d27a"} Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.552076 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m5vj" event={"ID":"ea745621-9b83-4af8-bb18-7fea76a4167d","Type":"ContainerStarted","Data":"c403a51331d766948e8f58288db55cf67728b559a06483a5d011cb949beab995"} Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.570811 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p2vmw"] Dec 09 11:54:14 crc kubenswrapper[4770]: W1209 11:54:14.593641 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b963017_2a12_4106_b6da_49b6725bfd9d.slice/crio-2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2 WatchSource:0}: Error finding container 2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2: Status 404 returned error can't find the container with id 2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2 Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.595473 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.618819 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-l9l6m"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.633386 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6b8c6ab-f03c-4927-8a49-c174a34e1787" (UID: "f6b8c6ab-f03c-4927-8a49-c174a34e1787"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.667200 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zjjm9"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.679413 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.707620 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.735867 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.736294 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.815224 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.862832 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config" (OuterVolumeSpecName: "config") pod "f6b8c6ab-f03c-4927-8a49-c174a34e1787" (UID: "f6b8c6ab-f03c-4927-8a49-c174a34e1787"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.865559 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6b8c6ab-f03c-4927-8a49-c174a34e1787" (UID: "f6b8c6ab-f03c-4927-8a49-c174a34e1787"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.882682 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.882714 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.895196 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6b8c6ab-f03c-4927-8a49-c174a34e1787" (UID: "f6b8c6ab-f03c-4927-8a49-c174a34e1787"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:14 crc kubenswrapper[4770]: I1209 11:54:14.986621 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b8c6ab-f03c-4927-8a49-c174a34e1787-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.158785 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.197736 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.207748 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bfd654465-rlbn9"] Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.430917 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" path="/var/lib/kubelet/pods/f6b8c6ab-f03c-4927-8a49-c174a34e1787/volumes" Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.544648 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:15 crc kubenswrapper[4770]: W1209 11:54:15.563133 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode52e1b12_1774_4832_a6fa_9d3c430d1408.slice/crio-fd27dac480a59d9ee2d3e0fb10cb1cb6905cdb47f941449b60f59b48deb49e85 WatchSource:0}: Error finding container fd27dac480a59d9ee2d3e0fb10cb1cb6905cdb47f941449b60f59b48deb49e85: Status 404 returned error can't find the container with id fd27dac480a59d9ee2d3e0fb10cb1cb6905cdb47f941449b60f59b48deb49e85 Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.608914 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-l9l6m" event={"ID":"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c","Type":"ContainerStarted","Data":"8d873c38c1b605a0973fce198ea3596a148416b4b267624233cff3d2a70e4864"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.608973 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-l9l6m" event={"ID":"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c","Type":"ContainerStarted","Data":"17bb94ba71992c019568903ebdc946c238ee8f3024f20134f9765962eae680a7"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.634151 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zjjm9" event={"ID":"f2433461-ff2f-4364-adca-d649cd8d68a9","Type":"ContainerStarted","Data":"a68ce4b2167d79aa35039d85fe012ba7aa3da1f20c383ba2dbf632b543474e4d"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.657945 4770 generic.go:334] "Generic (PLEG): container finished" podID="03c1adca-9233-4b79-a995-1d7e498d08a5" containerID="729643fcc7ca0a61bf79ca8c53aaef17a2f4bc4ca73bf42a91a86516bb8f6daa" exitCode=0 Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.658159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" event={"ID":"03c1adca-9233-4b79-a995-1d7e498d08a5","Type":"ContainerDied","Data":"729643fcc7ca0a61bf79ca8c53aaef17a2f4bc4ca73bf42a91a86516bb8f6daa"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.666939 4770 generic.go:334] "Generic (PLEG): container finished" podID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerID="ad64e843239020286eb2efab998bb5addb41d21dd4f5bdf8ac1e771dc03bd503" exitCode=0 Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.667221 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" event={"ID":"5124dab3-0ca0-40b9-beea-133de08b32bf","Type":"ContainerDied","Data":"ad64e843239020286eb2efab998bb5addb41d21dd4f5bdf8ac1e771dc03bd503"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.667278 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" event={"ID":"5124dab3-0ca0-40b9-beea-133de08b32bf","Type":"ContainerStarted","Data":"92dcabc76883fcb9d8aa824ce8ab15b77676a0c89b4b49ca9d713485c5f59713"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.699114 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wr87b" event={"ID":"289530c3-8968-4421-91b9-325d95b941c4","Type":"ContainerStarted","Data":"846a9a013e8d368afaaf333fa475f4a8d0eece651d21d6b459c099f7297229b4"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.701955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p2vmw" event={"ID":"3b963017-2a12-4106-b6da-49b6725bfd9d","Type":"ContainerStarted","Data":"2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.707806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"b10aa3a14105600acae388ed793d07042cdff68f5e1a6fb0546ab50d2924e520"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.720024 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerStarted","Data":"e46aac983b5d1aa36d90b9a9b6934390c07a772ae9ec61632100586b57e7cae2"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.729160 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerStarted","Data":"bc62fabba4d3dc8f1e3160e6d872a8fd89615e30ecd85bd0d62354448d75a559"} Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.755985 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-l9l6m" podStartSLOduration=3.755959624 podStartE2EDuration="3.755959624s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:15.748680729 +0000 UTC m=+1380.989439248" watchObservedRunningTime="2025-12-09 11:54:15.755959624 +0000 UTC m=+1380.996718143" Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.838853 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.898827 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wr87b" podStartSLOduration=3.898804991 podStartE2EDuration="3.898804991s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:15.896048041 +0000 UTC m=+1381.136806560" watchObservedRunningTime="2025-12-09 11:54:15.898804991 +0000 UTC m=+1381.139563510" Dec 09 11:54:15 crc kubenswrapper[4770]: I1209 11:54:15.962050 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.015602 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.442380 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.483315 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkxv7\" (UniqueName: \"kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7\") pod \"03c1adca-9233-4b79-a995-1d7e498d08a5\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.483425 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb\") pod \"03c1adca-9233-4b79-a995-1d7e498d08a5\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.483493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc\") pod \"03c1adca-9233-4b79-a995-1d7e498d08a5\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.483521 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb\") pod \"03c1adca-9233-4b79-a995-1d7e498d08a5\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.483669 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config\") pod \"03c1adca-9233-4b79-a995-1d7e498d08a5\" (UID: \"03c1adca-9233-4b79-a995-1d7e498d08a5\") " Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.529784 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03c1adca-9233-4b79-a995-1d7e498d08a5" (UID: "03c1adca-9233-4b79-a995-1d7e498d08a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.530653 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03c1adca-9233-4b79-a995-1d7e498d08a5" (UID: "03c1adca-9233-4b79-a995-1d7e498d08a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.535749 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7" (OuterVolumeSpecName: "kube-api-access-kkxv7") pod "03c1adca-9233-4b79-a995-1d7e498d08a5" (UID: "03c1adca-9233-4b79-a995-1d7e498d08a5"). InnerVolumeSpecName "kube-api-access-kkxv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.550437 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03c1adca-9233-4b79-a995-1d7e498d08a5" (UID: "03c1adca-9233-4b79-a995-1d7e498d08a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.558146 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config" (OuterVolumeSpecName: "config") pod "03c1adca-9233-4b79-a995-1d7e498d08a5" (UID: "03c1adca-9233-4b79-a995-1d7e498d08a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.587848 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.587893 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkxv7\" (UniqueName: \"kubernetes.io/projected/03c1adca-9233-4b79-a995-1d7e498d08a5-kube-api-access-kkxv7\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.587951 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.587962 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.587972 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03c1adca-9233-4b79-a995-1d7e498d08a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.768320 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.768495 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99559fbf5-kfgdk" event={"ID":"03c1adca-9233-4b79-a995-1d7e498d08a5","Type":"ContainerDied","Data":"21121f0cec7f4ee54be03c3911e5b8ec6d290d5a8c4ce1817d287efb4a80d27a"} Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.768562 4770 scope.go:117] "RemoveContainer" containerID="729643fcc7ca0a61bf79ca8c53aaef17a2f4bc4ca73bf42a91a86516bb8f6daa" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.774846 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" event={"ID":"5124dab3-0ca0-40b9-beea-133de08b32bf","Type":"ContainerStarted","Data":"82ee564cfdbaeec2bb8702efb46f9cac138e67a8f49e60bfcf68c03cdda57aed"} Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.774983 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.795144 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" podStartSLOduration=3.795122171 podStartE2EDuration="3.795122171s" podCreationTimestamp="2025-12-09 11:54:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:16.792437522 +0000 UTC m=+1382.033196041" watchObservedRunningTime="2025-12-09 11:54:16.795122171 +0000 UTC m=+1382.035880700" Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.797644 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerStarted","Data":"fd27dac480a59d9ee2d3e0fb10cb1cb6905cdb47f941449b60f59b48deb49e85"} Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.912093 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:16 crc kubenswrapper[4770]: I1209 11:54:16.920388 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-99559fbf5-kfgdk"] Dec 09 11:54:17 crc kubenswrapper[4770]: I1209 11:54:17.353553 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c1adca-9233-4b79-a995-1d7e498d08a5" path="/var/lib/kubelet/pods/03c1adca-9233-4b79-a995-1d7e498d08a5/volumes" Dec 09 11:54:18 crc kubenswrapper[4770]: I1209 11:54:18.831546 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerStarted","Data":"2fabc2876a4bc5a5c951fe8c4d0954bc1a3f02ab7ad9eba35ae7a642909f2e7d"} Dec 09 11:54:18 crc kubenswrapper[4770]: I1209 11:54:18.834429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerStarted","Data":"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb"} Dec 09 11:54:21 crc kubenswrapper[4770]: I1209 11:54:21.871771 4770 generic.go:334] "Generic (PLEG): container finished" podID="289530c3-8968-4421-91b9-325d95b941c4" containerID="846a9a013e8d368afaaf333fa475f4a8d0eece651d21d6b459c099f7297229b4" exitCode=0 Dec 09 11:54:21 crc kubenswrapper[4770]: I1209 11:54:21.871851 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wr87b" event={"ID":"289530c3-8968-4421-91b9-325d95b941c4","Type":"ContainerDied","Data":"846a9a013e8d368afaaf333fa475f4a8d0eece651d21d6b459c099f7297229b4"} Dec 09 11:54:23 crc kubenswrapper[4770]: I1209 11:54:23.626104 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:54:23 crc kubenswrapper[4770]: I1209 11:54:23.740563 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:54:23 crc kubenswrapper[4770]: I1209 11:54:23.740932 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" containerID="cri-o://dcfabf3d3df79b39f1a1ad99c77d0fba948ca668b9b09caaf8c8a7c3db3857b4" gracePeriod=10 Dec 09 11:54:23 crc kubenswrapper[4770]: I1209 11:54:23.912510 4770 generic.go:334] "Generic (PLEG): container finished" podID="ab1ee405-e664-46bc-a816-1526373ada15" containerID="dcfabf3d3df79b39f1a1ad99c77d0fba948ca668b9b09caaf8c8a7c3db3857b4" exitCode=0 Dec 09 11:54:23 crc kubenswrapper[4770]: I1209 11:54:23.912565 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" event={"ID":"ab1ee405-e664-46bc-a816-1526373ada15","Type":"ContainerDied","Data":"dcfabf3d3df79b39f1a1ad99c77d0fba948ca668b9b09caaf8c8a7c3db3857b4"} Dec 09 11:54:24 crc kubenswrapper[4770]: I1209 11:54:24.089254 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.460103 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559238 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559279 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx924\" (UniqueName: \"kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559384 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.559456 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys\") pod \"289530c3-8968-4421-91b9-325d95b941c4\" (UID: \"289530c3-8968-4421-91b9-325d95b941c4\") " Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.569450 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924" (OuterVolumeSpecName: "kube-api-access-qx924") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "kube-api-access-qx924". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.571075 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.581340 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts" (OuterVolumeSpecName: "scripts") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.592426 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.604292 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data" (OuterVolumeSpecName: "config-data") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.633025 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "289530c3-8968-4421-91b9-325d95b941c4" (UID: "289530c3-8968-4421-91b9-325d95b941c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661615 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661663 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661678 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661690 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx924\" (UniqueName: \"kubernetes.io/projected/289530c3-8968-4421-91b9-325d95b941c4-kube-api-access-qx924\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661705 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.661717 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/289530c3-8968-4421-91b9-325d95b941c4-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.942663 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wr87b" event={"ID":"289530c3-8968-4421-91b9-325d95b941c4","Type":"ContainerDied","Data":"ee1f899fe60cc1c9c5521e75450ed62da3beb63a31782c7b57617f9acd978c03"} Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.943022 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee1f899fe60cc1c9c5521e75450ed62da3beb63a31782c7b57617f9acd978c03" Dec 09 11:54:26 crc kubenswrapper[4770]: I1209 11:54:26.942732 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wr87b" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.761396 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wr87b"] Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.776841 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wr87b"] Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.858357 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jzfqf"] Dec 09 11:54:27 crc kubenswrapper[4770]: E1209 11:54:27.859016 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289530c3-8968-4421-91b9-325d95b941c4" containerName="keystone-bootstrap" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859038 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="289530c3-8968-4421-91b9-325d95b941c4" containerName="keystone-bootstrap" Dec 09 11:54:27 crc kubenswrapper[4770]: E1209 11:54:27.859058 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="init" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859066 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="init" Dec 09 11:54:27 crc kubenswrapper[4770]: E1209 11:54:27.859096 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c1adca-9233-4b79-a995-1d7e498d08a5" containerName="init" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859102 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c1adca-9233-4b79-a995-1d7e498d08a5" containerName="init" Dec 09 11:54:27 crc kubenswrapper[4770]: E1209 11:54:27.859117 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="dnsmasq-dns" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859126 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="dnsmasq-dns" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859318 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="289530c3-8968-4421-91b9-325d95b941c4" containerName="keystone-bootstrap" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859340 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b8c6ab-f03c-4927-8a49-c174a34e1787" containerName="dnsmasq-dns" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.859351 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c1adca-9233-4b79-a995-1d7e498d08a5" containerName="init" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.860040 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.862949 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.863171 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.863415 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sz4h5" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.863495 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.865143 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.868000 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jzfqf"] Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.998969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.999749 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.999847 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq9wt\" (UniqueName: \"kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:27 crc kubenswrapper[4770]: I1209 11:54:27.999942 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.000061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.000091 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq9wt\" (UniqueName: \"kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102348 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.102415 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.108448 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.109989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.112951 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.128134 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.130755 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.141117 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq9wt\" (UniqueName: \"kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt\") pod \"keystone-bootstrap-jzfqf\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:28 crc kubenswrapper[4770]: I1209 11:54:28.190570 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:29 crc kubenswrapper[4770]: I1209 11:54:29.352419 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="289530c3-8968-4421-91b9-325d95b941c4" path="/var/lib/kubelet/pods/289530c3-8968-4421-91b9-325d95b941c4/volumes" Dec 09 11:54:32 crc kubenswrapper[4770]: I1209 11:54:32.474306 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:54:32 crc kubenswrapper[4770]: I1209 11:54:32.474373 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:54:34 crc kubenswrapper[4770]: I1209 11:54:34.089505 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 09 11:54:36 crc kubenswrapper[4770]: E1209 11:54:36.114438 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16" Dec 09 11:54:36 crc kubenswrapper[4770]: E1209 11:54:36.115188 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbwlz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-p2vmw_openstack(3b963017-2a12-4106-b6da-49b6725bfd9d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:54:36 crc kubenswrapper[4770]: E1209 11:54:36.116513 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-p2vmw" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" Dec 09 11:54:36 crc kubenswrapper[4770]: E1209 11:54:36.779136 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5a548c25fe3d02f7a042cb0a6d28fc8039a34c4a3b3d07aadda4aba3a926e777" Dec 09 11:54:36 crc kubenswrapper[4770]: E1209 11:54:36.779364 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5a548c25fe3d02f7a042cb0a6d28fc8039a34c4a3b3d07aadda4aba3a926e777,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bbh54bh75h4h59dhdbh569h5fh5b4h666h597hd5hffh5c8h65h6fh64bh56bhb7h58bhdbh5b9h68bh5c4hdbh5fh656hc4h5c6h96h545h9cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qc2sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(14c7c388-f457-44e4-ae7a-22904d6aa33c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:54:36 crc kubenswrapper[4770]: I1209 11:54:36.866678 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.056153 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.056337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" event={"ID":"ab1ee405-e664-46bc-a816-1526373ada15","Type":"ContainerDied","Data":"433da4e8899de70be5ce6d63d2f0c0b86acaa6ff72561be5438920e6e91f87a1"} Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.056474 4770 scope.go:117] "RemoveContainer" containerID="dcfabf3d3df79b39f1a1ad99c77d0fba948ca668b9b09caaf8c8a7c3db3857b4" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.057014 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc\") pod \"ab1ee405-e664-46bc-a816-1526373ada15\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.057036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb\") pod \"ab1ee405-e664-46bc-a816-1526373ada15\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.057090 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb\") pod \"ab1ee405-e664-46bc-a816-1526373ada15\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.057179 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5wvx\" (UniqueName: \"kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx\") pod \"ab1ee405-e664-46bc-a816-1526373ada15\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.057299 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config\") pod \"ab1ee405-e664-46bc-a816-1526373ada15\" (UID: \"ab1ee405-e664-46bc-a816-1526373ada15\") " Dec 09 11:54:37 crc kubenswrapper[4770]: E1209 11:54:37.061180 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16\\\"\"" pod="openstack/barbican-db-sync-p2vmw" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.063366 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx" (OuterVolumeSpecName: "kube-api-access-d5wvx") pod "ab1ee405-e664-46bc-a816-1526373ada15" (UID: "ab1ee405-e664-46bc-a816-1526373ada15"). InnerVolumeSpecName "kube-api-access-d5wvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.118681 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab1ee405-e664-46bc-a816-1526373ada15" (UID: "ab1ee405-e664-46bc-a816-1526373ada15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.119831 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config" (OuterVolumeSpecName: "config") pod "ab1ee405-e664-46bc-a816-1526373ada15" (UID: "ab1ee405-e664-46bc-a816-1526373ada15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.121739 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab1ee405-e664-46bc-a816-1526373ada15" (UID: "ab1ee405-e664-46bc-a816-1526373ada15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.122411 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab1ee405-e664-46bc-a816-1526373ada15" (UID: "ab1ee405-e664-46bc-a816-1526373ada15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.160469 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.160507 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.160519 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.160530 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab1ee405-e664-46bc-a816-1526373ada15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.160544 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5wvx\" (UniqueName: \"kubernetes.io/projected/ab1ee405-e664-46bc-a816-1526373ada15-kube-api-access-d5wvx\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.394597 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.403269 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-klzb9"] Dec 09 11:54:37 crc kubenswrapper[4770]: I1209 11:54:37.970444 4770 scope.go:117] "RemoveContainer" containerID="d501d6b22d091e0904ae87a07713eca3b978c87e116a5d4755b1a061f6aa2702" Dec 09 11:54:38 crc kubenswrapper[4770]: E1209 11:54:38.004741 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49" Dec 09 11:54:38 crc kubenswrapper[4770]: E1209 11:54:38.005518 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xrvr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-4m5vj_openstack(ea745621-9b83-4af8-bb18-7fea76a4167d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:54:38 crc kubenswrapper[4770]: E1209 11:54:38.006846 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-4m5vj" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" Dec 09 11:54:38 crc kubenswrapper[4770]: E1209 11:54:38.100682 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49\\\"\"" pod="openstack/cinder-db-sync-4m5vj" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" Dec 09 11:54:38 crc kubenswrapper[4770]: I1209 11:54:38.462453 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jzfqf"] Dec 09 11:54:38 crc kubenswrapper[4770]: W1209 11:54:38.469887 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8299ef4_cea1_4ddc_b9a7_dd61fb6b9647.slice/crio-bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80 WatchSource:0}: Error finding container bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80: Status 404 returned error can't find the container with id bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80 Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.099298 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-klzb9" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.114967 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerStarted","Data":"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.115063 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-log" containerID="cri-o://57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" gracePeriod=30 Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.115105 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-httpd" containerID="cri-o://6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" gracePeriod=30 Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.118250 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zjjm9" event={"ID":"f2433461-ff2f-4364-adca-d649cd8d68a9","Type":"ContainerStarted","Data":"b398d9c6572a1c1d9989fa5b3cdc1de7aed57f3c833bda8fd8efb536cd6ce5db"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.123574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.123623 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.123637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.139714 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jzfqf" event={"ID":"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647","Type":"ContainerStarted","Data":"60819c01211cdf868787f2091ef5192fb54f532e2c708f8cbbacdddd49d63cda"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.139769 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jzfqf" event={"ID":"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647","Type":"ContainerStarted","Data":"bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.155234 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=27.155208417 podStartE2EDuration="27.155208417s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:39.144969507 +0000 UTC m=+1404.385728046" watchObservedRunningTime="2025-12-09 11:54:39.155208417 +0000 UTC m=+1404.395966946" Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.156628 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerStarted","Data":"28dbf913dc50f85ef74f8e63a3bc54e5e037d339134b47c1c460a199ab69d209"} Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.156855 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-log" containerID="cri-o://2fabc2876a4bc5a5c951fe8c4d0954bc1a3f02ab7ad9eba35ae7a642909f2e7d" gracePeriod=30 Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.157011 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-httpd" containerID="cri-o://28dbf913dc50f85ef74f8e63a3bc54e5e037d339134b47c1c460a199ab69d209" gracePeriod=30 Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.182746 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jzfqf" podStartSLOduration=12.182723856 podStartE2EDuration="12.182723856s" podCreationTimestamp="2025-12-09 11:54:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:39.165937649 +0000 UTC m=+1404.406696188" watchObservedRunningTime="2025-12-09 11:54:39.182723856 +0000 UTC m=+1404.423482375" Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.232141 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zjjm9" podStartSLOduration=4.152791643 podStartE2EDuration="26.23211137s" podCreationTimestamp="2025-12-09 11:54:13 +0000 UTC" firstStartedPulling="2025-12-09 11:54:14.688594952 +0000 UTC m=+1379.929353471" lastFinishedPulling="2025-12-09 11:54:36.767914689 +0000 UTC m=+1402.008673198" observedRunningTime="2025-12-09 11:54:39.193859098 +0000 UTC m=+1404.434617617" watchObservedRunningTime="2025-12-09 11:54:39.23211137 +0000 UTC m=+1404.472869889" Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.251110 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.251082701 podStartE2EDuration="27.251082701s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:39.238458551 +0000 UTC m=+1404.479217080" watchObservedRunningTime="2025-12-09 11:54:39.251082701 +0000 UTC m=+1404.491841220" Dec 09 11:54:39 crc kubenswrapper[4770]: I1209 11:54:39.360674 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab1ee405-e664-46bc-a816-1526373ada15" path="/var/lib/kubelet/pods/ab1ee405-e664-46bc-a816-1526373ada15/volumes" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.083463 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.173480 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.175961 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerStarted","Data":"1e31cdf21f8cff9f7bd69dcb8a17c433f4dede72f5120c87e25f7f8705b55889"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.181628 4770 generic.go:334] "Generic (PLEG): container finished" podID="93a65856-3106-400c-901c-623fbf408f48" containerID="28dbf913dc50f85ef74f8e63a3bc54e5e037d339134b47c1c460a199ab69d209" exitCode=0 Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.182803 4770 generic.go:334] "Generic (PLEG): container finished" podID="93a65856-3106-400c-901c-623fbf408f48" containerID="2fabc2876a4bc5a5c951fe8c4d0954bc1a3f02ab7ad9eba35ae7a642909f2e7d" exitCode=143 Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.182884 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerDied","Data":"28dbf913dc50f85ef74f8e63a3bc54e5e037d339134b47c1c460a199ab69d209"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.182936 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerDied","Data":"2fabc2876a4bc5a5c951fe8c4d0954bc1a3f02ab7ad9eba35ae7a642909f2e7d"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.186957 4770 generic.go:334] "Generic (PLEG): container finished" podID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerID="6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" exitCode=0 Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187002 4770 generic.go:334] "Generic (PLEG): container finished" podID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerID="57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" exitCode=143 Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187018 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187145 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerDied","Data":"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187190 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerDied","Data":"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e52e1b12-1774-4832-a6fa-9d3c430d1408","Type":"ContainerDied","Data":"fd27dac480a59d9ee2d3e0fb10cb1cb6905cdb47f941449b60f59b48deb49e85"} Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.187221 4770 scope.go:117] "RemoveContainer" containerID="6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246042 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246278 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246340 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246367 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjvr6\" (UniqueName: \"kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246480 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246566 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.246682 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts\") pod \"e52e1b12-1774-4832-a6fa-9d3c430d1408\" (UID: \"e52e1b12-1774-4832-a6fa-9d3c430d1408\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.247693 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs" (OuterVolumeSpecName: "logs") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.247785 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.248381 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.253716 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts" (OuterVolumeSpecName: "scripts") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.255557 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6" (OuterVolumeSpecName: "kube-api-access-kjvr6") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "kube-api-access-kjvr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.255707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.291083 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.324570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.327385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data" (OuterVolumeSpecName: "config-data") pod "e52e1b12-1774-4832-a6fa-9d3c430d1408" (UID: "e52e1b12-1774-4832-a6fa-9d3c430d1408"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350442 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350477 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350486 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjvr6\" (UniqueName: \"kubernetes.io/projected/e52e1b12-1774-4832-a6fa-9d3c430d1408-kube-api-access-kjvr6\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350497 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350505 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e52e1b12-1774-4832-a6fa-9d3c430d1408-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350529 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.350538 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52e1b12-1774-4832-a6fa-9d3c430d1408-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.374245 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.397592 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.404332 4770 scope.go:117] "RemoveContainer" containerID="57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.452372 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.467689 4770 scope.go:117] "RemoveContainer" containerID="6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.470330 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e\": container with ID starting with 6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e not found: ID does not exist" containerID="6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.470442 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e"} err="failed to get container status \"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e\": rpc error: code = NotFound desc = could not find container \"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e\": container with ID starting with 6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e not found: ID does not exist" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.470499 4770 scope.go:117] "RemoveContainer" containerID="57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.471262 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb\": container with ID starting with 57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb not found: ID does not exist" containerID="57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.471300 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb"} err="failed to get container status \"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb\": rpc error: code = NotFound desc = could not find container \"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb\": container with ID starting with 57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb not found: ID does not exist" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.471328 4770 scope.go:117] "RemoveContainer" containerID="6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.471805 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e"} err="failed to get container status \"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e\": rpc error: code = NotFound desc = could not find container \"6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e\": container with ID starting with 6a9aafe5c7b9fda28a42789bfa333dad92143d220fd4bbf45606aead1e0be86e not found: ID does not exist" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.471850 4770 scope.go:117] "RemoveContainer" containerID="57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.472141 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb"} err="failed to get container status \"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb\": rpc error: code = NotFound desc = could not find container \"57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb\": container with ID starting with 57888ae317dfea306b2f50078c2cb579df597cc7457e4d77b7e046d343cb91eb not found: ID does not exist" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.540390 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.553882 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.559938 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560207 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560261 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560352 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560428 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560487 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjxw\" (UniqueName: \"kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.560531 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts\") pod \"93a65856-3106-400c-901c-623fbf408f48\" (UID: \"93a65856-3106-400c-901c-623fbf408f48\") " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.561765 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs" (OuterVolumeSpecName: "logs") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.562540 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.563399 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.569423 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw" (OuterVolumeSpecName: "kube-api-access-2tjxw") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "kube-api-access-2tjxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.577795 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.578845 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.592671 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593342 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593369 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593393 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593419 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593427 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593444 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593452 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593473 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593504 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: E1209 11:54:40.593529 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="init" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593541 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="init" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593793 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab1ee405-e664-46bc-a816-1526373ada15" containerName="dnsmasq-dns" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593817 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593834 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593853 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" containerName="glance-log" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.593867 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a65856-3106-400c-901c-623fbf408f48" containerName="glance-httpd" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.595329 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.597875 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.599517 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.600176 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts" (OuterVolumeSpecName: "scripts") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.608476 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.666239 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667242 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667598 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667638 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667653 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/93a65856-3106-400c-901c-623fbf408f48-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667666 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667678 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjxw\" (UniqueName: \"kubernetes.io/projected/93a65856-3106-400c-901c-623fbf408f48-kube-api-access-2tjxw\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.667689 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.680046 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data" (OuterVolumeSpecName: "config-data") pod "93a65856-3106-400c-901c-623fbf408f48" (UID: "93a65856-3106-400c-901c-623fbf408f48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.700633 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769101 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769157 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769241 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769278 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769366 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2xzd\" (UniqueName: \"kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769399 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769601 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769940 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.769975 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93a65856-3106-400c-901c-623fbf408f48-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872138 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872218 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872250 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872287 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872343 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2xzd\" (UniqueName: \"kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872409 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872454 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.872540 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.876938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.877035 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.880346 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.881823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.882686 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.891127 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.891145 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.895557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2xzd\" (UniqueName: \"kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:40 crc kubenswrapper[4770]: I1209 11:54:40.926136 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.201430 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.201419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"93a65856-3106-400c-901c-623fbf408f48","Type":"ContainerDied","Data":"bc62fabba4d3dc8f1e3160e6d872a8fd89615e30ecd85bd0d62354448d75a559"} Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.202879 4770 scope.go:117] "RemoveContainer" containerID="28dbf913dc50f85ef74f8e63a3bc54e5e037d339134b47c1c460a199ab69d209" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.227411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.243088 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.261734 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.264376 4770 scope.go:117] "RemoveContainer" containerID="2fabc2876a4bc5a5c951fe8c4d0954bc1a3f02ab7ad9eba35ae7a642909f2e7d" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.285698 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.287500 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.292337 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.301474 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.304562 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.362239 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a65856-3106-400c-901c-623fbf408f48" path="/var/lib/kubelet/pods/93a65856-3106-400c-901c-623fbf408f48/volumes" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.362902 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52e1b12-1774-4832-a6fa-9d3c430d1408" path="/var/lib/kubelet/pods/e52e1b12-1774-4832-a6fa-9d3c430d1408/volumes" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401189 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401298 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401381 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7gpd\" (UniqueName: \"kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401415 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401493 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.401541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.508886 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.508990 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509099 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7gpd\" (UniqueName: \"kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509123 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509366 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509390 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.509861 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.510693 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.515400 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.521685 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.521887 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.522508 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.522730 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.532778 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7gpd\" (UniqueName: \"kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.551528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " pod="openstack/glance-default-external-api-0" Dec 09 11:54:41 crc kubenswrapper[4770]: I1209 11:54:41.610324 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:54:42 crc kubenswrapper[4770]: I1209 11:54:42.103738 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:54:42 crc kubenswrapper[4770]: W1209 11:54:42.152142 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81346ca2_bec1_4d0c_900a_2af788284bc7.slice/crio-fedd2f06e5ed46971a39c6dc0e03d77a5424ea16cebd322156baaf4adf7555c3 WatchSource:0}: Error finding container fedd2f06e5ed46971a39c6dc0e03d77a5424ea16cebd322156baaf4adf7555c3: Status 404 returned error can't find the container with id fedd2f06e5ed46971a39c6dc0e03d77a5424ea16cebd322156baaf4adf7555c3 Dec 09 11:54:42 crc kubenswrapper[4770]: I1209 11:54:42.233992 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerStarted","Data":"fedd2f06e5ed46971a39c6dc0e03d77a5424ea16cebd322156baaf4adf7555c3"} Dec 09 11:54:42 crc kubenswrapper[4770]: I1209 11:54:42.238733 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364"} Dec 09 11:54:43 crc kubenswrapper[4770]: I1209 11:54:43.175619 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:54:43 crc kubenswrapper[4770]: W1209 11:54:43.192570 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4043b25c_3d5f_4f4d_8eea_9990caab521c.slice/crio-9014350d818bc33f317c92132e8085363c6a71a504debb4b2d6e1454db4996bc WatchSource:0}: Error finding container 9014350d818bc33f317c92132e8085363c6a71a504debb4b2d6e1454db4996bc: Status 404 returned error can't find the container with id 9014350d818bc33f317c92132e8085363c6a71a504debb4b2d6e1454db4996bc Dec 09 11:54:43 crc kubenswrapper[4770]: I1209 11:54:43.366817 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab"} Dec 09 11:54:43 crc kubenswrapper[4770]: I1209 11:54:43.366857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621"} Dec 09 11:54:43 crc kubenswrapper[4770]: I1209 11:54:43.369687 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerStarted","Data":"b0b0312f945dfc048ae1a1bc134866f989faa293193ee9b494cd073af664a917"} Dec 09 11:54:43 crc kubenswrapper[4770]: I1209 11:54:43.397142 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerStarted","Data":"9014350d818bc33f317c92132e8085363c6a71a504debb4b2d6e1454db4996bc"} Dec 09 11:54:44 crc kubenswrapper[4770]: I1209 11:54:44.414955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerStarted","Data":"7f7eaccb9a37b54a7a21a7a819223a1106f81f476a8f56d4926ef19ebddbb708"} Dec 09 11:54:45 crc kubenswrapper[4770]: I1209 11:54:45.426412 4770 generic.go:334] "Generic (PLEG): container finished" podID="b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" containerID="60819c01211cdf868787f2091ef5192fb54f532e2c708f8cbbacdddd49d63cda" exitCode=0 Dec 09 11:54:45 crc kubenswrapper[4770]: I1209 11:54:45.426524 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jzfqf" event={"ID":"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647","Type":"ContainerDied","Data":"60819c01211cdf868787f2091ef5192fb54f532e2c708f8cbbacdddd49d63cda"} Dec 09 11:54:45 crc kubenswrapper[4770]: I1209 11:54:45.428855 4770 generic.go:334] "Generic (PLEG): container finished" podID="f2433461-ff2f-4364-adca-d649cd8d68a9" containerID="b398d9c6572a1c1d9989fa5b3cdc1de7aed57f3c833bda8fd8efb536cd6ce5db" exitCode=0 Dec 09 11:54:45 crc kubenswrapper[4770]: I1209 11:54:45.428886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zjjm9" event={"ID":"f2433461-ff2f-4364-adca-d649cd8d68a9","Type":"ContainerDied","Data":"b398d9c6572a1c1d9989fa5b3cdc1de7aed57f3c833bda8fd8efb536cd6ce5db"} Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.446249 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zjjm9" event={"ID":"f2433461-ff2f-4364-adca-d649cd8d68a9","Type":"ContainerDied","Data":"a68ce4b2167d79aa35039d85fe012ba7aa3da1f20c383ba2dbf632b543474e4d"} Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.446771 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68ce4b2167d79aa35039d85fe012ba7aa3da1f20c383ba2dbf632b543474e4d" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.448412 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jzfqf" event={"ID":"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647","Type":"ContainerDied","Data":"bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80"} Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.448451 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bba3b4206af6c697e8dfb75fe2c7a63689af1732588ce267beadbdc451526f80" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.449772 4770 generic.go:334] "Generic (PLEG): container finished" podID="07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" containerID="8d873c38c1b605a0973fce198ea3596a148416b4b267624233cff3d2a70e4864" exitCode=0 Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.449801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-l9l6m" event={"ID":"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c","Type":"ContainerDied","Data":"8d873c38c1b605a0973fce198ea3596a148416b4b267624233cff3d2a70e4864"} Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.490854 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.524687 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577057 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq9wt\" (UniqueName: \"kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577127 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data\") pod \"f2433461-ff2f-4364-adca-d649cd8d68a9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577768 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle\") pod \"f2433461-ff2f-4364-adca-d649cd8d68a9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577869 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577897 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5qng\" (UniqueName: \"kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng\") pod \"f2433461-ff2f-4364-adca-d649cd8d68a9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.577971 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.578035 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.578078 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts\") pod \"f2433461-ff2f-4364-adca-d649cd8d68a9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.578106 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs\") pod \"f2433461-ff2f-4364-adca-d649cd8d68a9\" (UID: \"f2433461-ff2f-4364-adca-d649cd8d68a9\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.578153 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys\") pod \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\" (UID: \"b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647\") " Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.583365 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs" (OuterVolumeSpecName: "logs") pod "f2433461-ff2f-4364-adca-d649cd8d68a9" (UID: "f2433461-ff2f-4364-adca-d649cd8d68a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.585988 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng" (OuterVolumeSpecName: "kube-api-access-m5qng") pod "f2433461-ff2f-4364-adca-d649cd8d68a9" (UID: "f2433461-ff2f-4364-adca-d649cd8d68a9"). InnerVolumeSpecName "kube-api-access-m5qng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.590501 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts" (OuterVolumeSpecName: "scripts") pod "f2433461-ff2f-4364-adca-d649cd8d68a9" (UID: "f2433461-ff2f-4364-adca-d649cd8d68a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.592177 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt" (OuterVolumeSpecName: "kube-api-access-nq9wt") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "kube-api-access-nq9wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.592495 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts" (OuterVolumeSpecName: "scripts") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.600284 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.603984 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.656353 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data" (OuterVolumeSpecName: "config-data") pod "f2433461-ff2f-4364-adca-d649cd8d68a9" (UID: "f2433461-ff2f-4364-adca-d649cd8d68a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.674100 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681100 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq9wt\" (UniqueName: \"kubernetes.io/projected/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-kube-api-access-nq9wt\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681142 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681153 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5qng\" (UniqueName: \"kubernetes.io/projected/f2433461-ff2f-4364-adca-d649cd8d68a9-kube-api-access-m5qng\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681161 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681171 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681179 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681188 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681196 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2433461-ff2f-4364-adca-d649cd8d68a9-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681204 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.681936 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data" (OuterVolumeSpecName: "config-data") pod "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" (UID: "b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.685491 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2433461-ff2f-4364-adca-d649cd8d68a9" (UID: "f2433461-ff2f-4364-adca-d649cd8d68a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.792153 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:47 crc kubenswrapper[4770]: I1209 11:54:47.792209 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2433461-ff2f-4364-adca-d649cd8d68a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.463183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerStarted","Data":"34e7162111d82c416b36d7cac62876c550b4c2a95c5dd62c86e63169addfa649"} Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.470351 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerStarted","Data":"334ccf338ffd966eccc95375cc7b064fdbb4ecd1a1b5f283b0cb6961e3e4d9aa"} Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.478857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1"} Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.480867 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jzfqf" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.483124 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zjjm9" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.483243 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerStarted","Data":"668171f0a1f1cde0fe4f4fe471f6859804937404f8461d50cc1a370f3b4c29dd"} Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.512516 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.512494616 podStartE2EDuration="8.512494616s" podCreationTimestamp="2025-12-09 11:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:48.491023491 +0000 UTC m=+1413.731782010" watchObservedRunningTime="2025-12-09 11:54:48.512494616 +0000 UTC m=+1413.753253135" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.521340 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.52132448 podStartE2EDuration="7.52132448s" podCreationTimestamp="2025-12-09 11:54:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:48.519125094 +0000 UTC m=+1413.759883613" watchObservedRunningTime="2025-12-09 11:54:48.52132448 +0000 UTC m=+1413.762082999" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.685613 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:54:48 crc kubenswrapper[4770]: E1209 11:54:48.686055 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" containerName="keystone-bootstrap" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.686074 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" containerName="keystone-bootstrap" Dec 09 11:54:48 crc kubenswrapper[4770]: E1209 11:54:48.686092 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2433461-ff2f-4364-adca-d649cd8d68a9" containerName="placement-db-sync" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.686098 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2433461-ff2f-4364-adca-d649cd8d68a9" containerName="placement-db-sync" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.686277 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2433461-ff2f-4364-adca-d649cd8d68a9" containerName="placement-db-sync" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.686299 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" containerName="keystone-bootstrap" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.687020 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.696810 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.697043 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.697245 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.697384 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.697486 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sz4h5" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.698154 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.709045 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.743947 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.749070 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.752155 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5tc9t" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.752315 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.752421 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.755632 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.755824 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.759617 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.811951 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812034 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812112 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812141 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812179 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzwjx\" (UniqueName: \"kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812222 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812257 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.812290 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914491 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914521 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914631 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914662 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914678 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914710 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914745 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914766 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwjx\" (UniqueName: \"kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914812 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914838 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdjqb\" (UniqueName: \"kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.914872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.921868 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.922077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.922086 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.922575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.925014 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.926797 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.931199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:48 crc kubenswrapper[4770]: I1209 11:54:48.936308 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwjx\" (UniqueName: \"kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx\") pod \"keystone-5b4585bb4d-2j2md\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.016868 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017048 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdjqb\" (UniqueName: \"kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017169 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017636 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.017715 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.018322 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.021621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.022385 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.023056 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.023081 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.032519 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.033872 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.033983 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdjqb\" (UniqueName: \"kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb\") pod \"placement-574ccd49f8-67m27\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:49 crc kubenswrapper[4770]: I1209 11:54:49.084154 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.228404 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.228954 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.263211 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.288482 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.514708 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.514788 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.611104 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.611162 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.646601 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 11:54:51 crc kubenswrapper[4770]: I1209 11:54:51.670261 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 11:54:52 crc kubenswrapper[4770]: I1209 11:54:52.523765 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 11:54:52 crc kubenswrapper[4770]: I1209 11:54:52.524134 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.044426 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.143582 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hfpg\" (UniqueName: \"kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg\") pod \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.144088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config\") pod \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.144310 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle\") pod \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\" (UID: \"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c\") " Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.178762 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg" (OuterVolumeSpecName: "kube-api-access-6hfpg") pod "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" (UID: "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c"). InnerVolumeSpecName "kube-api-access-6hfpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.187503 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" (UID: "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.221062 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config" (OuterVolumeSpecName: "config") pod "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" (UID: "07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.248070 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.248112 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hfpg\" (UniqueName: \"kubernetes.io/projected/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-kube-api-access-6hfpg\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.248126 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.455568 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.533769 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b4585bb4d-2j2md" event={"ID":"b306554d-ccb5-4db1-8462-fef6aba58f57","Type":"ContainerStarted","Data":"2f5d6fb1acb5dbd5aca8d99eb0093122d83a493146fd4dc86ff66d2986c069d2"} Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.536752 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-l9l6m" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.537508 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-l9l6m" event={"ID":"07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c","Type":"ContainerDied","Data":"17bb94ba71992c019568903ebdc946c238ee8f3024f20134f9765962eae680a7"} Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.537530 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17bb94ba71992c019568903ebdc946c238ee8f3024f20134f9765962eae680a7" Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.636797 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:54:53 crc kubenswrapper[4770]: I1209 11:54:53.920282 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.242404 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:54:54 crc kubenswrapper[4770]: E1209 11:54:54.243130 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" containerName="neutron-db-sync" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.243147 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" containerName="neutron-db-sync" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.243331 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" containerName="neutron-db-sync" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.244300 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.258850 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.284708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzst9\" (UniqueName: \"kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.284844 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.284955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.284992 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.285089 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.313972 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.315887 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.324427 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.324653 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.324963 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.325101 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6tnc9" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.339122 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386644 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386702 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzst9\" (UniqueName: \"kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386747 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386766 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386784 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386813 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386834 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386895 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk475\" (UniqueName: \"kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.386928 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.387846 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.388079 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.388153 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.388749 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.407329 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzst9\" (UniqueName: \"kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9\") pod \"dnsmasq-dns-5846f59699-7kf24\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.488035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.488795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.488940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.489063 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.489177 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk475\" (UniqueName: \"kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.499370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.499633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.500927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.509034 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk475\" (UniqueName: \"kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.524954 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config\") pod \"neutron-8455957bd-72jgv\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.547430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerStarted","Data":"26689815601d539a1e5a64d638a9b7e9d573190bd6a436f8545ce167279468c4"} Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.575037 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:54 crc kubenswrapper[4770]: I1209 11:54:54.647299 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.098979 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:54:55 crc kubenswrapper[4770]: W1209 11:54:55.128149 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3a81174_6679_4c08_b966_4094af54c7d8.slice/crio-e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391 WatchSource:0}: Error finding container e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391: Status 404 returned error can't find the container with id e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391 Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.192997 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.193303 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 11:54:55 crc kubenswrapper[4770]: W1209 11:54:55.317752 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e501958_bf84_4a46_a0ee_526ebccdaf60.slice/crio-f10b6f305df51099d0f045b5dc06277278500f84deb63a306cc51a46c77f2f2a WatchSource:0}: Error finding container f10b6f305df51099d0f045b5dc06277278500f84deb63a306cc51a46c77f2f2a: Status 404 returned error can't find the container with id f10b6f305df51099d0f045b5dc06277278500f84deb63a306cc51a46c77f2f2a Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.337120 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.567083 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerStarted","Data":"f10b6f305df51099d0f045b5dc06277278500f84deb63a306cc51a46c77f2f2a"} Dec 09 11:54:55 crc kubenswrapper[4770]: I1209 11:54:55.569747 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerStarted","Data":"e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391"} Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.455793 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.759797 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.768218 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.772003 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.772305 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.777605 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.841626 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.841691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.841980 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h5gw\" (UniqueName: \"kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.842128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.842215 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.842453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.842618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944115 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944278 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h5gw\" (UniqueName: \"kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944339 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944382 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.944426 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.950930 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.951011 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.954746 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.956511 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.956630 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.960968 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:56 crc kubenswrapper[4770]: I1209 11:54:56.966107 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h5gw\" (UniqueName: \"kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw\") pod \"neutron-6dc5b689b7-qgl92\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.102356 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.612833 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerStarted","Data":"297de169838c4bac5ceb5ffe3cb876b689e8b9332529573ff7910aa430b83e44"} Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.649617 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerStarted","Data":"bd89441d5382fde55ff03f8f5a849f41ce9ccbe18c8191c1ab37523eab113b45"} Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.668746 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b4585bb4d-2j2md" event={"ID":"b306554d-ccb5-4db1-8462-fef6aba58f57","Type":"ContainerStarted","Data":"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919"} Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.670010 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.714176 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerStarted","Data":"4ba87b24c19c1120bbcabbace77c866035e530eaf47c87572eba2d7e47526454"} Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.789863 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b4585bb4d-2j2md" podStartSLOduration=9.789837076 podStartE2EDuration="9.789837076s" podCreationTimestamp="2025-12-09 11:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:57.732413228 +0000 UTC m=+1422.973171747" watchObservedRunningTime="2025-12-09 11:54:57.789837076 +0000 UTC m=+1423.030595595" Dec 09 11:54:57 crc kubenswrapper[4770]: I1209 11:54:57.798571 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:54:57 crc kubenswrapper[4770]: W1209 11:54:57.800185 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ab03763_40e0_4e54_bb2a_7ac9de8c5184.slice/crio-b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d WatchSource:0}: Error finding container b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d: Status 404 returned error can't find the container with id b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.746853 4770 generic.go:334] "Generic (PLEG): container finished" podID="c3a81174-6679-4c08-b966-4094af54c7d8" containerID="297de169838c4bac5ceb5ffe3cb876b689e8b9332529573ff7910aa430b83e44" exitCode=0 Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.747575 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerDied","Data":"297de169838c4bac5ceb5ffe3cb876b689e8b9332529573ff7910aa430b83e44"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.769588 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.776829 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerStarted","Data":"54e36573ebf61b52f64abf81a2d9c40f6e426c644c97d07e3601d928f6cb4c31"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.777231 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.789331 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerStarted","Data":"b505b07675385a501eefd16b47033f1d0cdb0b120345d67e47e86e33ef0f2548"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.790390 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.790440 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.806797 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerStarted","Data":"85925850bf61e5c8f64a81b9f9e5037d619b24f1d0912e5127b4a2874bb652ee"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.806854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerStarted","Data":"fd8527bc2e276f42630b31d0e4c2d54c775ef3cdb370d0ca709fc11bfc05e843"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.806866 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerStarted","Data":"b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.807561 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.815078 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8455957bd-72jgv" podStartSLOduration=4.815057488 podStartE2EDuration="4.815057488s" podCreationTimestamp="2025-12-09 11:54:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:58.799960565 +0000 UTC m=+1424.040719094" watchObservedRunningTime="2025-12-09 11:54:58.815057488 +0000 UTC m=+1424.055815997" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.820476 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p2vmw" event={"ID":"3b963017-2a12-4106-b6da-49b6725bfd9d","Type":"ContainerStarted","Data":"74aa53e345d67730abe1cb4d00a4a413ff9c3de93027ad63aa017d74368659fe"} Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.842632 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-574ccd49f8-67m27" podStartSLOduration=10.842609308 podStartE2EDuration="10.842609308s" podCreationTimestamp="2025-12-09 11:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:58.83483764 +0000 UTC m=+1424.075596179" watchObservedRunningTime="2025-12-09 11:54:58.842609308 +0000 UTC m=+1424.083367827" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.879585 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-p2vmw" podStartSLOduration=3.6283047330000002 podStartE2EDuration="46.879561436s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="2025-12-09 11:54:14.621326914 +0000 UTC m=+1379.862085433" lastFinishedPulling="2025-12-09 11:54:57.872583617 +0000 UTC m=+1423.113342136" observedRunningTime="2025-12-09 11:54:58.857837044 +0000 UTC m=+1424.098595563" watchObservedRunningTime="2025-12-09 11:54:58.879561436 +0000 UTC m=+1424.120319955" Dec 09 11:54:58 crc kubenswrapper[4770]: I1209 11:54:58.894134 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6dc5b689b7-qgl92" podStartSLOduration=2.894105455 podStartE2EDuration="2.894105455s" podCreationTimestamp="2025-12-09 11:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:58.890278808 +0000 UTC m=+1424.131037357" watchObservedRunningTime="2025-12-09 11:54:58.894105455 +0000 UTC m=+1424.134863974" Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.834849 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerStarted","Data":"12e1f31a90f3e82dc874c093f9ae7defb528a89431960cfdaa426ccc760094c3"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.835498 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.865293 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5846f59699-7kf24" podStartSLOduration=5.865272625 podStartE2EDuration="5.865272625s" podCreationTimestamp="2025-12-09 11:54:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:54:59.857857317 +0000 UTC m=+1425.098615836" watchObservedRunningTime="2025-12-09 11:54:59.865272625 +0000 UTC m=+1425.106031144" Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.883688 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.883736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.883746 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.883755 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.889005 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m5vj" event={"ID":"ea745621-9b83-4af8-bb18-7fea76a4167d","Type":"ContainerStarted","Data":"6909006db23528ef5901641d836bab057983dc7c9e64e0113cbd39ca6d6fab1d"} Dec 09 11:54:59 crc kubenswrapper[4770]: I1209 11:54:59.926159 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4m5vj" podStartSLOduration=4.004078995 podStartE2EDuration="47.926136981s" podCreationTimestamp="2025-12-09 11:54:12 +0000 UTC" firstStartedPulling="2025-12-09 11:54:14.200721854 +0000 UTC m=+1379.441480373" lastFinishedPulling="2025-12-09 11:54:58.12277984 +0000 UTC m=+1423.363538359" observedRunningTime="2025-12-09 11:54:59.919471151 +0000 UTC m=+1425.160229670" watchObservedRunningTime="2025-12-09 11:54:59.926136981 +0000 UTC m=+1425.166895500" Dec 09 11:55:00 crc kubenswrapper[4770]: I1209 11:55:00.906867 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6"} Dec 09 11:55:00 crc kubenswrapper[4770]: I1209 11:55:00.907274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerStarted","Data":"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d"} Dec 09 11:55:00 crc kubenswrapper[4770]: I1209 11:55:00.960564 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=69.954250404 podStartE2EDuration="1m52.960547077s" podCreationTimestamp="2025-12-09 11:53:08 +0000 UTC" firstStartedPulling="2025-12-09 11:54:14.911163974 +0000 UTC m=+1380.151922493" lastFinishedPulling="2025-12-09 11:54:57.917460647 +0000 UTC m=+1423.158219166" observedRunningTime="2025-12-09 11:55:00.952321478 +0000 UTC m=+1426.193080007" watchObservedRunningTime="2025-12-09 11:55:00.960547077 +0000 UTC m=+1426.201305596" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.249247 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.272693 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.287400 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.299410 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.299829 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471344 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6995z\" (UniqueName: \"kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471439 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471521 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471590 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.471752 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.573894 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574071 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574176 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6995z\" (UniqueName: \"kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574218 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574252 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574282 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.574794 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.575017 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.575228 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.575341 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.576530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.594180 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6995z\" (UniqueName: \"kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z\") pod \"dnsmasq-dns-7bb67c87c9-bzrd8\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.626389 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.921190 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b963017-2a12-4106-b6da-49b6725bfd9d" containerID="74aa53e345d67730abe1cb4d00a4a413ff9c3de93027ad63aa017d74368659fe" exitCode=0 Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.921404 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5846f59699-7kf24" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="dnsmasq-dns" containerID="cri-o://12e1f31a90f3e82dc874c093f9ae7defb528a89431960cfdaa426ccc760094c3" gracePeriod=10 Dec 09 11:55:01 crc kubenswrapper[4770]: I1209 11:55:01.921530 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p2vmw" event={"ID":"3b963017-2a12-4106-b6da-49b6725bfd9d","Type":"ContainerDied","Data":"74aa53e345d67730abe1cb4d00a4a413ff9c3de93027ad63aa017d74368659fe"} Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.473525 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.473937 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.473992 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.475026 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.475106 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f" gracePeriod=600 Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.933070 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f" exitCode=0 Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.933152 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f"} Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.933195 4770 scope.go:117] "RemoveContainer" containerID="107d94b1c663bb7c1f2192c993f8baba84266be8138fe32dc0a89e9051aeba2c" Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.935244 4770 generic.go:334] "Generic (PLEG): container finished" podID="c3a81174-6679-4c08-b966-4094af54c7d8" containerID="12e1f31a90f3e82dc874c093f9ae7defb528a89431960cfdaa426ccc760094c3" exitCode=0 Dec 09 11:55:02 crc kubenswrapper[4770]: I1209 11:55:02.935456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerDied","Data":"12e1f31a90f3e82dc874c093f9ae7defb528a89431960cfdaa426ccc760094c3"} Dec 09 11:55:04 crc kubenswrapper[4770]: I1209 11:55:04.576979 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5846f59699-7kf24" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: connect: connection refused" Dec 09 11:55:06 crc kubenswrapper[4770]: I1209 11:55:06.978957 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p2vmw" event={"ID":"3b963017-2a12-4106-b6da-49b6725bfd9d","Type":"ContainerDied","Data":"2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2"} Dec 09 11:55:06 crc kubenswrapper[4770]: I1209 11:55:06.979481 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2770bea506a3c9af8c8a0274dea40a303ddac149b963a37d869deb27c38972c2" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.008913 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.065571 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle\") pod \"3b963017-2a12-4106-b6da-49b6725bfd9d\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.065749 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbwlz\" (UniqueName: \"kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz\") pod \"3b963017-2a12-4106-b6da-49b6725bfd9d\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.065968 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data\") pod \"3b963017-2a12-4106-b6da-49b6725bfd9d\" (UID: \"3b963017-2a12-4106-b6da-49b6725bfd9d\") " Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.074664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3b963017-2a12-4106-b6da-49b6725bfd9d" (UID: "3b963017-2a12-4106-b6da-49b6725bfd9d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.079143 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz" (OuterVolumeSpecName: "kube-api-access-kbwlz") pod "3b963017-2a12-4106-b6da-49b6725bfd9d" (UID: "3b963017-2a12-4106-b6da-49b6725bfd9d"). InnerVolumeSpecName "kube-api-access-kbwlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.110158 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b963017-2a12-4106-b6da-49b6725bfd9d" (UID: "3b963017-2a12-4106-b6da-49b6725bfd9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.168460 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.168510 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b963017-2a12-4106-b6da-49b6725bfd9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:07 crc kubenswrapper[4770]: I1209 11:55:07.168522 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbwlz\" (UniqueName: \"kubernetes.io/projected/3b963017-2a12-4106-b6da-49b6725bfd9d-kube-api-access-kbwlz\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:07.999776 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p2vmw" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:07.999886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5846f59699-7kf24" event={"ID":"c3a81174-6679-4c08-b966-4094af54c7d8","Type":"ContainerDied","Data":"e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391"} Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.000742 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e75d2fc4f01f9015dbe16cddccbbb9c7b2ca7cb60b2d9892d28ad9fa62c99391" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.017933 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.107658 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzst9\" (UniqueName: \"kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9\") pod \"c3a81174-6679-4c08-b966-4094af54c7d8\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.111119 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb\") pod \"c3a81174-6679-4c08-b966-4094af54c7d8\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.111168 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config\") pod \"c3a81174-6679-4c08-b966-4094af54c7d8\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.111276 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb\") pod \"c3a81174-6679-4c08-b966-4094af54c7d8\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.111389 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc\") pod \"c3a81174-6679-4c08-b966-4094af54c7d8\" (UID: \"c3a81174-6679-4c08-b966-4094af54c7d8\") " Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.113761 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9" (OuterVolumeSpecName: "kube-api-access-tzst9") pod "c3a81174-6679-4c08-b966-4094af54c7d8" (UID: "c3a81174-6679-4c08-b966-4094af54c7d8"). InnerVolumeSpecName "kube-api-access-tzst9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.165537 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3a81174-6679-4c08-b966-4094af54c7d8" (UID: "c3a81174-6679-4c08-b966-4094af54c7d8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.168500 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3a81174-6679-4c08-b966-4094af54c7d8" (UID: "c3a81174-6679-4c08-b966-4094af54c7d8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.177309 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3a81174-6679-4c08-b966-4094af54c7d8" (UID: "c3a81174-6679-4c08-b966-4094af54c7d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.181232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config" (OuterVolumeSpecName: "config") pod "c3a81174-6679-4c08-b966-4094af54c7d8" (UID: "c3a81174-6679-4c08-b966-4094af54c7d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.214320 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.214355 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzst9\" (UniqueName: \"kubernetes.io/projected/c3a81174-6679-4c08-b966-4094af54c7d8-kube-api-access-tzst9\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.214369 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.214380 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.214390 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3a81174-6679-4c08-b966-4094af54c7d8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.326648 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:55:08 crc kubenswrapper[4770]: E1209 11:55:08.327071 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="dnsmasq-dns" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.327087 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="dnsmasq-dns" Dec 09 11:55:08 crc kubenswrapper[4770]: E1209 11:55:08.327104 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="init" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.327110 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="init" Dec 09 11:55:08 crc kubenswrapper[4770]: E1209 11:55:08.327140 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" containerName="barbican-db-sync" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.327146 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" containerName="barbican-db-sync" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.327319 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" containerName="dnsmasq-dns" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.327335 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" containerName="barbican-db-sync" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.328490 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.330855 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.331623 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.335490 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qkbpv" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.353814 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.392349 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.419823 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.419917 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.419956 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gggtc\" (UniqueName: \"kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.420060 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.420128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: W1209 11:55:08.420581 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod002c4504_ccc8_4722_8d92_5255c2ade13a.slice/crio-acd350bdb515724438e8a4d24f0fbed50492e388bc30f335cf1fca26e57493fd WatchSource:0}: Error finding container acd350bdb515724438e8a4d24f0fbed50492e388bc30f335cf1fca26e57493fd: Status 404 returned error can't find the container with id acd350bdb515724438e8a4d24f0fbed50492e388bc30f335cf1fca26e57493fd Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.464833 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.467532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.472923 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.494504 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523214 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523272 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523318 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gggtc\" (UniqueName: \"kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523412 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vz5x\" (UniqueName: \"kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523557 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523638 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.523679 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.525668 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.526037 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.546125 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.562042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.568872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.572755 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gggtc\" (UniqueName: \"kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc\") pod \"barbican-worker-66bfc44b69-qfbmc\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.583643 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.590190 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625502 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625576 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7vpm\" (UniqueName: \"kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625673 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625718 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625752 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vz5x\" (UniqueName: \"kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625838 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.625845 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.626879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.626983 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.627014 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.632218 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.633034 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.633170 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.649993 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.651860 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.652907 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.656277 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.668019 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vz5x\" (UniqueName: \"kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x\") pod \"barbican-keystone-listener-764f86c974-ns5jl\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.670283 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.709283 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729372 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729481 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729516 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8djt2\" (UniqueName: \"kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729563 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7vpm\" (UniqueName: \"kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729598 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729624 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729648 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729682 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729713 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.729737 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.730557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.731641 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.732931 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.733426 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.735192 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.754231 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7vpm\" (UniqueName: \"kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm\") pod \"dnsmasq-dns-54c4dfcffc-9nrwp\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832189 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832280 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8djt2\" (UniqueName: \"kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832404 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832448 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832473 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.832727 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.837264 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.841944 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.843426 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.859452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8djt2\" (UniqueName: \"kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2\") pod \"barbican-api-6dcc946d68-b2gb6\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:08 crc kubenswrapper[4770]: E1209 11:55:08.866568 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.879257 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:55:08 crc kubenswrapper[4770]: I1209 11:55:08.931594 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.069402 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.100686 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerStarted","Data":"fa1c1004130919ce9c120230332027edd32a27025efbe2c594c15d1d0a1c4eee"} Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.101173 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="ceilometer-notification-agent" containerID="cri-o://1e31cdf21f8cff9f7bd69dcb8a17c433f4dede72f5120c87e25f7f8705b55889" gracePeriod=30 Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.101506 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.101880 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="proxy-httpd" containerID="cri-o://fa1c1004130919ce9c120230332027edd32a27025efbe2c594c15d1d0a1c4eee" gracePeriod=30 Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.101996 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="sg-core" containerID="cri-o://668171f0a1f1cde0fe4f4fe471f6859804937404f8461d50cc1a370f3b4c29dd" gracePeriod=30 Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.112930 4770 generic.go:334] "Generic (PLEG): container finished" podID="002c4504-ccc8-4722-8d92-5255c2ade13a" containerID="0ceedf9014d9bf3043a0b18debf44ff42db54e4689e9f791122cc5530547eaae" exitCode=0 Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.113014 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" event={"ID":"002c4504-ccc8-4722-8d92-5255c2ade13a","Type":"ContainerDied","Data":"0ceedf9014d9bf3043a0b18debf44ff42db54e4689e9f791122cc5530547eaae"} Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.113074 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" event={"ID":"002c4504-ccc8-4722-8d92-5255c2ade13a","Type":"ContainerStarted","Data":"acd350bdb515724438e8a4d24f0fbed50492e388bc30f335cf1fca26e57493fd"} Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.171467 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5846f59699-7kf24" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.173030 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118"} Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.369697 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.377681 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5846f59699-7kf24"] Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.383835 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:55:09 crc kubenswrapper[4770]: W1209 11:55:09.393371 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod568ef04e_032c_40f4_97b9_b21483281a4f.slice/crio-0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f WatchSource:0}: Error finding container 0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f: Status 404 returned error can't find the container with id 0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.417242 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.591870 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:55:09 crc kubenswrapper[4770]: W1209 11:55:09.616765 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f7b544a_d135_44fc_9fda_3938369f661f.slice/crio-5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9 WatchSource:0}: Error finding container 5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9: Status 404 returned error can't find the container with id 5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9 Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.632658 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.683821 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6995z\" (UniqueName: \"kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.683889 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.684006 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.684054 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.684125 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.684261 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb\") pod \"002c4504-ccc8-4722-8d92-5255c2ade13a\" (UID: \"002c4504-ccc8-4722-8d92-5255c2ade13a\") " Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.690392 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z" (OuterVolumeSpecName: "kube-api-access-6995z") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "kube-api-access-6995z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.708180 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.709335 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.711289 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.711807 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.715980 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config" (OuterVolumeSpecName: "config") pod "002c4504-ccc8-4722-8d92-5255c2ade13a" (UID: "002c4504-ccc8-4722-8d92-5255c2ade13a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785787 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6995z\" (UniqueName: \"kubernetes.io/projected/002c4504-ccc8-4722-8d92-5255c2ade13a-kube-api-access-6995z\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785821 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785834 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785848 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785861 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.785872 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/002c4504-ccc8-4722-8d92-5255c2ade13a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:09 crc kubenswrapper[4770]: W1209 11:55:09.791766 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280114f2_3fd3_4c5b_bada_a047a5553bbc.slice/crio-e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d WatchSource:0}: Error finding container e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d: Status 404 returned error can't find the container with id e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.793684 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:09 crc kubenswrapper[4770]: I1209 11:55:09.912466 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:09 crc kubenswrapper[4770]: W1209 11:55:09.921054 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b273e04_19aa_4d8e_b82c_95f822c25677.slice/crio-d29ee20737c507128588a8d583c6927dbeb76a672fe68aaac20c80559b30faa2 WatchSource:0}: Error finding container d29ee20737c507128588a8d583c6927dbeb76a672fe68aaac20c80559b30faa2: Status 404 returned error can't find the container with id d29ee20737c507128588a8d583c6927dbeb76a672fe68aaac20c80559b30faa2 Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.192287 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.192312 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb67c87c9-bzrd8" event={"ID":"002c4504-ccc8-4722-8d92-5255c2ade13a","Type":"ContainerDied","Data":"acd350bdb515724438e8a4d24f0fbed50492e388bc30f335cf1fca26e57493fd"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.193139 4770 scope.go:117] "RemoveContainer" containerID="0ceedf9014d9bf3043a0b18debf44ff42db54e4689e9f791122cc5530547eaae" Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.202241 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerStarted","Data":"da61d729c0b6b1e02f1c0a6ae649a12798627a0b0d85b763d039b8fe64855854"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.202293 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerStarted","Data":"e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.207273 4770 generic.go:334] "Generic (PLEG): container finished" podID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerID="668171f0a1f1cde0fe4f4fe471f6859804937404f8461d50cc1a370f3b4c29dd" exitCode=2 Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.207354 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerDied","Data":"668171f0a1f1cde0fe4f4fe471f6859804937404f8461d50cc1a370f3b4c29dd"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.210629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerStarted","Data":"0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.213380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerStarted","Data":"5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.215739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerStarted","Data":"8acac0c355671ed1bf61265f9f271acc9e87c858a0e6cd69df8c66b09d2912b3"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.215779 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerStarted","Data":"d29ee20737c507128588a8d583c6927dbeb76a672fe68aaac20c80559b30faa2"} Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.313127 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:10 crc kubenswrapper[4770]: I1209 11:55:10.334353 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bb67c87c9-bzrd8"] Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.234764 4770 generic.go:334] "Generic (PLEG): container finished" podID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerID="da61d729c0b6b1e02f1c0a6ae649a12798627a0b0d85b763d039b8fe64855854" exitCode=0 Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.235144 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerDied","Data":"da61d729c0b6b1e02f1c0a6ae649a12798627a0b0d85b763d039b8fe64855854"} Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.243719 4770 generic.go:334] "Generic (PLEG): container finished" podID="ea745621-9b83-4af8-bb18-7fea76a4167d" containerID="6909006db23528ef5901641d836bab057983dc7c9e64e0113cbd39ca6d6fab1d" exitCode=0 Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.243770 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m5vj" event={"ID":"ea745621-9b83-4af8-bb18-7fea76a4167d","Type":"ContainerDied","Data":"6909006db23528ef5901641d836bab057983dc7c9e64e0113cbd39ca6d6fab1d"} Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.367661 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="002c4504-ccc8-4722-8d92-5255c2ade13a" path="/var/lib/kubelet/pods/002c4504-ccc8-4722-8d92-5255c2ade13a/volumes" Dec 09 11:55:11 crc kubenswrapper[4770]: I1209 11:55:11.368337 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a81174-6679-4c08-b966-4094af54c7d8" path="/var/lib/kubelet/pods/c3a81174-6679-4c08-b966-4094af54c7d8/volumes" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.290007 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerStarted","Data":"b6ff8a331f1648332a1647a4235baa0db21ba20309d42c80141d918f7421332b"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.291811 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.291921 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.310608 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerStarted","Data":"948f5d8ee8bd59bbc4c4f416e35f1ff0d5497a183addd81d6baefcacf2368a39"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.311592 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.331547 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6dcc946d68-b2gb6" podStartSLOduration=4.331526238 podStartE2EDuration="4.331526238s" podCreationTimestamp="2025-12-09 11:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:12.320603061 +0000 UTC m=+1437.561361580" watchObservedRunningTime="2025-12-09 11:55:12.331526238 +0000 UTC m=+1437.572284757" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.344111 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerStarted","Data":"bf8ad42774200d477849ea268cd29b0c65b5c006ef52af1b6710676014a7e842"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.344175 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerStarted","Data":"e77c216dc47b6c7c9a614efb4c3cf8153cbd683bc99a04574e90b93b09aa208a"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.359088 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" podStartSLOduration=4.359075088 podStartE2EDuration="4.359075088s" podCreationTimestamp="2025-12-09 11:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:12.356200515 +0000 UTC m=+1437.596959044" watchObservedRunningTime="2025-12-09 11:55:12.359075088 +0000 UTC m=+1437.599833607" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.366243 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerStarted","Data":"3390dfb8cbb1f0c6bd1a75c816ce715ca36a3164d095376f2127e25ddb826b67"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.366314 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerStarted","Data":"c5ec8fa6eed14138a68a6b229d128065bd09d1ea7c60899215996437c5d9b880"} Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.411365 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-66bfc44b69-qfbmc" podStartSLOduration=2.520652386 podStartE2EDuration="4.411338125s" podCreationTimestamp="2025-12-09 11:55:08 +0000 UTC" firstStartedPulling="2025-12-09 11:55:09.417009612 +0000 UTC m=+1434.657768121" lastFinishedPulling="2025-12-09 11:55:11.307695341 +0000 UTC m=+1436.548453860" observedRunningTime="2025-12-09 11:55:12.402309596 +0000 UTC m=+1437.643068125" watchObservedRunningTime="2025-12-09 11:55:12.411338125 +0000 UTC m=+1437.652096634" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.413367 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" podStartSLOduration=2.706937887 podStartE2EDuration="4.413351286s" podCreationTimestamp="2025-12-09 11:55:08 +0000 UTC" firstStartedPulling="2025-12-09 11:55:09.623631359 +0000 UTC m=+1434.864389868" lastFinishedPulling="2025-12-09 11:55:11.330044748 +0000 UTC m=+1436.570803267" observedRunningTime="2025-12-09 11:55:12.383366915 +0000 UTC m=+1437.624125434" watchObservedRunningTime="2025-12-09 11:55:12.413351286 +0000 UTC m=+1437.654109815" Dec 09 11:55:12 crc kubenswrapper[4770]: I1209 11:55:12.888090 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.011856 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012011 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012097 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012142 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrvr4\" (UniqueName: \"kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012403 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts\") pod \"ea745621-9b83-4af8-bb18-7fea76a4167d\" (UID: \"ea745621-9b83-4af8-bb18-7fea76a4167d\") " Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.012966 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea745621-9b83-4af8-bb18-7fea76a4167d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.031616 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.040669 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4" (OuterVolumeSpecName: "kube-api-access-xrvr4") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "kube-api-access-xrvr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.048206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts" (OuterVolumeSpecName: "scripts") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.056720 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:55:13 crc kubenswrapper[4770]: E1209 11:55:13.057369 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" containerName="cinder-db-sync" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.057387 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" containerName="cinder-db-sync" Dec 09 11:55:13 crc kubenswrapper[4770]: E1209 11:55:13.057422 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="002c4504-ccc8-4722-8d92-5255c2ade13a" containerName="init" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.057430 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="002c4504-ccc8-4722-8d92-5255c2ade13a" containerName="init" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.057676 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" containerName="cinder-db-sync" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.057699 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="002c4504-ccc8-4722-8d92-5255c2ade13a" containerName="init" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.059129 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.070542 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.070728 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.077991 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.108047 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114083 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114155 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114233 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114291 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvl48\" (UniqueName: \"kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114316 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114361 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114449 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114465 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114477 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrvr4\" (UniqueName: \"kubernetes.io/projected/ea745621-9b83-4af8-bb18-7fea76a4167d-kube-api-access-xrvr4\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.114491 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.184158 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data" (OuterVolumeSpecName: "config-data") pod "ea745621-9b83-4af8-bb18-7fea76a4167d" (UID: "ea745621-9b83-4af8-bb18-7fea76a4167d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217115 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvl48\" (UniqueName: \"kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217280 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217337 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217387 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217476 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea745621-9b83-4af8-bb18-7fea76a4167d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.217872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.222939 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.225480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.226675 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.227293 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.227772 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.240733 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvl48\" (UniqueName: \"kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48\") pod \"barbican-api-5674985874-bhmwx\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.377419 4770 generic.go:334] "Generic (PLEG): container finished" podID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerID="1e31cdf21f8cff9f7bd69dcb8a17c433f4dede72f5120c87e25f7f8705b55889" exitCode=0 Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.377534 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerDied","Data":"1e31cdf21f8cff9f7bd69dcb8a17c433f4dede72f5120c87e25f7f8705b55889"} Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.381166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m5vj" event={"ID":"ea745621-9b83-4af8-bb18-7fea76a4167d","Type":"ContainerDied","Data":"c403a51331d766948e8f58288db55cf67728b559a06483a5d011cb949beab995"} Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.381242 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c403a51331d766948e8f58288db55cf67728b559a06483a5d011cb949beab995" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.381288 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m5vj" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.428386 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.678399 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.683696 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.686856 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.687477 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pxvc8" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.687703 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.687846 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.710005 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735256 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735323 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735351 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735415 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpmg7\" (UniqueName: \"kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735472 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.735513 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.824427 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839273 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839351 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839390 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839454 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpmg7\" (UniqueName: \"kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839523 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.839572 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:13 crc kubenswrapper[4770]: I1209 11:55:13.843797 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.144611 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.145255 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.153536 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.163808 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.265721 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.292028 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.315414 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.331952 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpmg7\" (UniqueName: \"kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7\") pod \"cinder-scheduler-0\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.333442 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.338156 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.349447 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.356448 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.366834 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.388125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395020 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395153 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395261 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395465 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395612 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrkc4\" (UniqueName: \"kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395698 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395780 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.395981 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.396086 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.396205 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b5vp\" (UniqueName: \"kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.396416 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.498885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.498976 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499011 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b5vp\" (UniqueName: \"kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499087 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499167 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499227 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499261 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499306 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499386 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrkc4\" (UniqueName: \"kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499434 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.499455 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.501462 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.503320 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.504224 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.505766 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.506721 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.507438 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.509177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.511591 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.516580 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.519499 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.547505 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.551669 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.566063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b5vp\" (UniqueName: \"kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp\") pod \"cinder-api-0\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.576598 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrkc4\" (UniqueName: \"kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4\") pod \"dnsmasq-dns-6b4f5fc4f-2lgx7\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.690584 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.721475 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:14 crc kubenswrapper[4770]: I1209 11:55:14.989784 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:15 crc kubenswrapper[4770]: W1209 11:55:15.027090 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb774e6dd_e8ac_4992_9129_a7fb80adff38.slice/crio-7b9ff7fb8261e52bb8b502861ce5de31d4913954df25bd6b5de93b3b2e152647 WatchSource:0}: Error finding container 7b9ff7fb8261e52bb8b502861ce5de31d4913954df25bd6b5de93b3b2e152647: Status 404 returned error can't find the container with id 7b9ff7fb8261e52bb8b502861ce5de31d4913954df25bd6b5de93b3b2e152647 Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.359098 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.428666 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerStarted","Data":"7b9ff7fb8261e52bb8b502861ce5de31d4913954df25bd6b5de93b3b2e152647"} Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.430932 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" event={"ID":"a21d330f-a9b8-49b7-a035-83c9215ced97","Type":"ContainerStarted","Data":"d4d34001c16ef9075ef6bb2cf1967166be3051826e2fc29771a0d9154eafb82f"} Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.445153 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="dnsmasq-dns" containerID="cri-o://948f5d8ee8bd59bbc4c4f416e35f1ff0d5497a183addd81d6baefcacf2368a39" gracePeriod=10 Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.445244 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerStarted","Data":"9b65050fb96dad2b68de130b2e74efdc4a249da8425436217ee12e2727d54d98"} Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.445292 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerStarted","Data":"0ed885e5102a41a0b21a56c6ea93cb492355c4740fed766c91160632cf4c5c43"} Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.445305 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerStarted","Data":"24d4877fa872b6053bef5de4f870b65764d0ce5e9ed33b03b1a6a39665590b34"} Dec 09 11:55:15 crc kubenswrapper[4770]: I1209 11:55:15.453105 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.458039 4770 generic.go:334] "Generic (PLEG): container finished" podID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerID="948f5d8ee8bd59bbc4c4f416e35f1ff0d5497a183addd81d6baefcacf2368a39" exitCode=0 Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.458147 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerDied","Data":"948f5d8ee8bd59bbc4c4f416e35f1ff0d5497a183addd81d6baefcacf2368a39"} Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.458730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" event={"ID":"280114f2-3fd3-4c5b-bada-a047a5553bbc","Type":"ContainerDied","Data":"e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d"} Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.458769 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9bb0d6a681e4a7553d0aa200bf4d11bf41276f2b3652b8e465aee472533361d" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.460780 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerStarted","Data":"5a789b701154ca6d4b9e11a7e645c955a505070b9e86f50663425d84778a4c56"} Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.463111 4770 generic.go:334] "Generic (PLEG): container finished" podID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerID="13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9" exitCode=0 Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.464629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" event={"ID":"a21d330f-a9b8-49b7-a035-83c9215ced97","Type":"ContainerDied","Data":"13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9"} Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.464680 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.464716 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.465561 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.498982 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5674985874-bhmwx" podStartSLOduration=3.498952357 podStartE2EDuration="3.498952357s" podCreationTimestamp="2025-12-09 11:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:16.496004033 +0000 UTC m=+1441.736762552" watchObservedRunningTime="2025-12-09 11:55:16.498952357 +0000 UTC m=+1441.739710876" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.569746 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.569827 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.569879 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7vpm\" (UniqueName: \"kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.570036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.570194 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.570255 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.614266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm" (OuterVolumeSpecName: "kube-api-access-q7vpm") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "kube-api-access-q7vpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.661451 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.678499 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.678839 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") pod \"280114f2-3fd3-4c5b-bada-a047a5553bbc\" (UID: \"280114f2-3fd3-4c5b-bada-a047a5553bbc\") " Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.679744 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.679773 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7vpm\" (UniqueName: \"kubernetes.io/projected/280114f2-3fd3-4c5b-bada-a047a5553bbc-kube-api-access-q7vpm\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:16 crc kubenswrapper[4770]: W1209 11:55:16.679850 4770 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/280114f2-3fd3-4c5b-bada-a047a5553bbc/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.679862 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.685418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config" (OuterVolumeSpecName: "config") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.693032 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.699056 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "280114f2-3fd3-4c5b-bada-a047a5553bbc" (UID: "280114f2-3fd3-4c5b-bada-a047a5553bbc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.781741 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.781781 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.781795 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:16 crc kubenswrapper[4770]: I1209 11:55:16.781806 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/280114f2-3fd3-4c5b-bada-a047a5553bbc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.497173 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerStarted","Data":"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366"} Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.507023 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" event={"ID":"a21d330f-a9b8-49b7-a035-83c9215ced97","Type":"ContainerStarted","Data":"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e"} Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.507088 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4dfcffc-9nrwp" Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.538572 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" podStartSLOduration=4.538549464 podStartE2EDuration="4.538549464s" podCreationTimestamp="2025-12-09 11:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:17.536514192 +0000 UTC m=+1442.777272711" watchObservedRunningTime="2025-12-09 11:55:17.538549464 +0000 UTC m=+1442.779307983" Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.561534 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.571280 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c4dfcffc-9nrwp"] Dec 09 11:55:17 crc kubenswrapper[4770]: I1209 11:55:17.869518 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:18 crc kubenswrapper[4770]: I1209 11:55:18.517632 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerStarted","Data":"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f"} Dec 09 11:55:18 crc kubenswrapper[4770]: I1209 11:55:18.518071 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 11:55:18 crc kubenswrapper[4770]: I1209 11:55:18.520687 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerStarted","Data":"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0"} Dec 09 11:55:18 crc kubenswrapper[4770]: I1209 11:55:18.521163 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:18 crc kubenswrapper[4770]: I1209 11:55:18.567992 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.567948453 podStartE2EDuration="4.567948453s" podCreationTimestamp="2025-12-09 11:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:18.557096867 +0000 UTC m=+1443.797855396" watchObservedRunningTime="2025-12-09 11:55:18.567948453 +0000 UTC m=+1443.808706972" Dec 09 11:55:19 crc kubenswrapper[4770]: I1209 11:55:19.356567 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" path="/var/lib/kubelet/pods/280114f2-3fd3-4c5b-bada-a047a5553bbc/volumes" Dec 09 11:55:19 crc kubenswrapper[4770]: I1209 11:55:19.532365 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerStarted","Data":"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd"} Dec 09 11:55:19 crc kubenswrapper[4770]: I1209 11:55:19.532593 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api-log" containerID="cri-o://0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" gracePeriod=30 Dec 09 11:55:19 crc kubenswrapper[4770]: I1209 11:55:19.532678 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api" containerID="cri-o://42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" gracePeriod=30 Dec 09 11:55:19 crc kubenswrapper[4770]: I1209 11:55:19.573357 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.679922044 podStartE2EDuration="6.573325431s" podCreationTimestamp="2025-12-09 11:55:13 +0000 UTC" firstStartedPulling="2025-12-09 11:55:15.047283536 +0000 UTC m=+1440.288042055" lastFinishedPulling="2025-12-09 11:55:16.940686923 +0000 UTC m=+1442.181445442" observedRunningTime="2025-12-09 11:55:19.569483764 +0000 UTC m=+1444.810242293" watchObservedRunningTime="2025-12-09 11:55:19.573325431 +0000 UTC m=+1444.814083950" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.520879 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.543696 4770 generic.go:334] "Generic (PLEG): container finished" podID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerID="42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" exitCode=0 Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.543735 4770 generic.go:334] "Generic (PLEG): container finished" podID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerID="0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" exitCode=143 Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.544947 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.545419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerDied","Data":"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f"} Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.545451 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerDied","Data":"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366"} Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.545461 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0887b5b7-fa46-4386-b6b5-f26153d98f52","Type":"ContainerDied","Data":"5a789b701154ca6d4b9e11a7e645c955a505070b9e86f50663425d84778a4c56"} Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.545479 4770 scope.go:117] "RemoveContainer" containerID="42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554525 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554572 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554614 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554643 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554694 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b5vp\" (UniqueName: \"kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554743 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.554887 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom\") pod \"0887b5b7-fa46-4386-b6b5-f26153d98f52\" (UID: \"0887b5b7-fa46-4386-b6b5-f26153d98f52\") " Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.557035 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.558781 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0887b5b7-fa46-4386-b6b5-f26153d98f52-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.563332 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs" (OuterVolumeSpecName: "logs") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.569085 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts" (OuterVolumeSpecName: "scripts") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.573949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp" (OuterVolumeSpecName: "kube-api-access-2b5vp") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "kube-api-access-2b5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.579695 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.597180 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.599769 4770 scope.go:117] "RemoveContainer" containerID="0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.660413 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b5vp\" (UniqueName: \"kubernetes.io/projected/0887b5b7-fa46-4386-b6b5-f26153d98f52-kube-api-access-2b5vp\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.660655 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.660666 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.660675 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.660687 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0887b5b7-fa46-4386-b6b5-f26153d98f52-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.668625 4770 scope.go:117] "RemoveContainer" containerID="42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.668854 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data" (OuterVolumeSpecName: "config-data") pod "0887b5b7-fa46-4386-b6b5-f26153d98f52" (UID: "0887b5b7-fa46-4386-b6b5-f26153d98f52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.676687 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f\": container with ID starting with 42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f not found: ID does not exist" containerID="42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.676743 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f"} err="failed to get container status \"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f\": rpc error: code = NotFound desc = could not find container \"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f\": container with ID starting with 42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f not found: ID does not exist" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.676778 4770 scope.go:117] "RemoveContainer" containerID="0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.683397 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366\": container with ID starting with 0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366 not found: ID does not exist" containerID="0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.683468 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366"} err="failed to get container status \"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366\": rpc error: code = NotFound desc = could not find container \"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366\": container with ID starting with 0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366 not found: ID does not exist" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.683495 4770 scope.go:117] "RemoveContainer" containerID="42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.693435 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f"} err="failed to get container status \"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f\": rpc error: code = NotFound desc = could not find container \"42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f\": container with ID starting with 42f8473a70a8c252bf056e600d54c29f9cda2ae251e916a82f22b1836fb94d8f not found: ID does not exist" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.693499 4770 scope.go:117] "RemoveContainer" containerID="0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.694035 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366"} err="failed to get container status \"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366\": rpc error: code = NotFound desc = could not find container \"0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366\": container with ID starting with 0b6743b9b4655acc9681c299637445982f48540746df88b8ff9a228796ce0366 not found: ID does not exist" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.765603 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0887b5b7-fa46-4386-b6b5-f26153d98f52-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.884604 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.896663 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.924769 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.925640 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="dnsmasq-dns" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.925740 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="dnsmasq-dns" Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.925811 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api-log" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.925876 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api-log" Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.926010 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="init" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.926080 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="init" Dec 09 11:55:20 crc kubenswrapper[4770]: E1209 11:55:20.926172 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.926280 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.926642 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api-log" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.926726 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="280114f2-3fd3-4c5b-bada-a047a5553bbc" containerName="dnsmasq-dns" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.926817 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" containerName="cinder-api" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.929313 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.936018 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.936098 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.936806 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.947226 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w9mq\" (UniqueName: \"kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975573 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975628 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975657 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975730 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975786 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975814 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:20 crc kubenswrapper[4770]: I1209 11:55:20.975844 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.078539 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w9mq\" (UniqueName: \"kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.078877 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.079106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.079223 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.079352 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.079549 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.079874 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.080147 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.080285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.080406 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.090167 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.090334 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.090554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.091553 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.098641 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.106438 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.117749 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.213272 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w9mq\" (UniqueName: \"kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq\") pod \"cinder-api-0\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.264857 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.353807 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0887b5b7-fa46-4386-b6b5-f26153d98f52" path="/var/lib/kubelet/pods/0887b5b7-fa46-4386-b6b5-f26153d98f52/volumes" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.501474 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.614311 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:21 crc kubenswrapper[4770]: I1209 11:55:21.822470 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:55:21 crc kubenswrapper[4770]: W1209 11:55:21.828549 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c0a1baa_962d_4dfc_891e_3563feff00bf.slice/crio-714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74 WatchSource:0}: Error finding container 714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74: Status 404 returned error can't find the container with id 714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74 Dec 09 11:55:22 crc kubenswrapper[4770]: I1209 11:55:22.411518 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:55:22 crc kubenswrapper[4770]: I1209 11:55:22.597711 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:55:22 crc kubenswrapper[4770]: I1209 11:55:22.600326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerStarted","Data":"714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74"} Dec 09 11:55:23 crc kubenswrapper[4770]: I1209 11:55:23.434723 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:23 crc kubenswrapper[4770]: I1209 11:55:23.638788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerStarted","Data":"7bc283a32d2b16f74f1cf489c744ab75e9f7ab17baf7e2b5ac315e76c1b74b42"} Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.143218 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.144781 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.147793 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.148629 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.149521 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wwzpk" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.162658 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.221806 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.221961 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.222141 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.222204 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2nw4\" (UniqueName: \"kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.324191 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.324697 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.324740 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2nw4\" (UniqueName: \"kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.324808 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.329833 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.334628 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.334677 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.510204 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.533699 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2nw4\" (UniqueName: \"kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4\") pod \"openstackclient\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.662468 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.669502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerStarted","Data":"28cd7242a9c9640679276453e171ec22b26b2720ffe571f2f85b629ee46a5e06"} Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.669717 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.693727 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.717235 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.717209574 podStartE2EDuration="4.717209574s" podCreationTimestamp="2025-12-09 11:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:24.707173729 +0000 UTC m=+1449.947932248" watchObservedRunningTime="2025-12-09 11:55:24.717209574 +0000 UTC m=+1449.957968093" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.794589 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.826378 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.826686 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="dnsmasq-dns" containerID="cri-o://82ee564cfdbaeec2bb8702efb46f9cac138e67a8f49e60bfcf68c03cdda57aed" gracePeriod=10 Dec 09 11:55:24 crc kubenswrapper[4770]: I1209 11:55:24.899377 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.022916 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.694552 4770 generic.go:334] "Generic (PLEG): container finished" podID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerID="82ee564cfdbaeec2bb8702efb46f9cac138e67a8f49e60bfcf68c03cdda57aed" exitCode=0 Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.695395 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="cinder-scheduler" containerID="cri-o://4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" gracePeriod=30 Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.695975 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" event={"ID":"5124dab3-0ca0-40b9-beea-133de08b32bf","Type":"ContainerDied","Data":"82ee564cfdbaeec2bb8702efb46f9cac138e67a8f49e60bfcf68c03cdda57aed"} Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.701371 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="probe" containerID="cri-o://4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" gracePeriod=30 Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.831703 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 11:55:25 crc kubenswrapper[4770]: I1209 11:55:25.933502 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.094489 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b26x4\" (UniqueName: \"kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.094917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.094972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.095031 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.095184 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.128136 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4" (OuterVolumeSpecName: "kube-api-access-b26x4") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "kube-api-access-b26x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.177614 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config" (OuterVolumeSpecName: "config") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.196871 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.196972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") pod \"5124dab3-0ca0-40b9-beea-133de08b32bf\" (UID: \"5124dab3-0ca0-40b9-beea-133de08b32bf\") " Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.197462 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.197482 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b26x4\" (UniqueName: \"kubernetes.io/projected/5124dab3-0ca0-40b9-beea-133de08b32bf-kube-api-access-b26x4\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:26 crc kubenswrapper[4770]: W1209 11:55:26.197656 4770 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5124dab3-0ca0-40b9-beea-133de08b32bf/volumes/kubernetes.io~configmap/ovsdbserver-sb Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.197664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.203737 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.227099 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5124dab3-0ca0-40b9-beea-133de08b32bf" (UID: "5124dab3-0ca0-40b9-beea-133de08b32bf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.299200 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.299233 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.299243 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5124dab3-0ca0-40b9-beea-133de08b32bf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.674596 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.719021 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f6188401-2bb4-4ddc-a097-6b99f99df9e8","Type":"ContainerStarted","Data":"bf2b794cf42430c02ee120fad97763b7074bccc4e5b34d34d55e8836ab287611"} Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.722773 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" event={"ID":"5124dab3-0ca0-40b9-beea-133de08b32bf","Type":"ContainerDied","Data":"92dcabc76883fcb9d8aa824ce8ab15b77676a0c89b4b49ca9d713485c5f59713"} Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.722848 4770 scope.go:117] "RemoveContainer" containerID="82ee564cfdbaeec2bb8702efb46f9cac138e67a8f49e60bfcf68c03cdda57aed" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.722862 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f84f7cd9-k9wdm" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.770048 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.777400 4770 scope.go:117] "RemoveContainer" containerID="ad64e843239020286eb2efab998bb5addb41d21dd4f5bdf8ac1e771dc03bd503" Dec 09 11:55:26 crc kubenswrapper[4770]: I1209 11:55:26.781856 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67f84f7cd9-k9wdm"] Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.143477 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.216410 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.229216 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8455957bd-72jgv" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-api" containerID="cri-o://bd89441d5382fde55ff03f8f5a849f41ce9ccbe18c8191c1ab37523eab113b45" gracePeriod=30 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.229388 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8455957bd-72jgv" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-httpd" containerID="cri-o://54e36573ebf61b52f64abf81a2d9c40f6e426c644c97d07e3601d928f6cb4c31" gracePeriod=30 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.398580 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" path="/var/lib/kubelet/pods/5124dab3-0ca0-40b9-beea-133de08b32bf/volumes" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.433587 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.469758 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.546599 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.546661 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpmg7\" (UniqueName: \"kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.546710 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.546797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.547040 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.547750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.547812 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts\") pod \"b774e6dd-e8ac-4992-9129-a7fb80adff38\" (UID: \"b774e6dd-e8ac-4992-9129-a7fb80adff38\") " Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.548504 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b774e6dd-e8ac-4992-9129-a7fb80adff38-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.583815 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts" (OuterVolumeSpecName: "scripts") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.588134 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.613436 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.613721 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6dcc946d68-b2gb6" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api-log" containerID="cri-o://8acac0c355671ed1bf61265f9f271acc9e87c858a0e6cd69df8c66b09d2912b3" gracePeriod=30 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.614313 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6dcc946d68-b2gb6" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api" containerID="cri-o://b6ff8a331f1648332a1647a4235baa0db21ba20309d42c80141d918f7421332b" gracePeriod=30 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.646533 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7" (OuterVolumeSpecName: "kube-api-access-xpmg7") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "kube-api-access-xpmg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.648498 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.650067 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.650101 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.650113 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.650126 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpmg7\" (UniqueName: \"kubernetes.io/projected/b774e6dd-e8ac-4992-9129-a7fb80adff38-kube-api-access-xpmg7\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.711465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data" (OuterVolumeSpecName: "config-data") pod "b774e6dd-e8ac-4992-9129-a7fb80adff38" (UID: "b774e6dd-e8ac-4992-9129-a7fb80adff38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.751914 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b774e6dd-e8ac-4992-9129-a7fb80adff38-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.781030 4770 generic.go:334] "Generic (PLEG): container finished" podID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerID="54e36573ebf61b52f64abf81a2d9c40f6e426c644c97d07e3601d928f6cb4c31" exitCode=0 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.781312 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerDied","Data":"54e36573ebf61b52f64abf81a2d9c40f6e426c644c97d07e3601d928f6cb4c31"} Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791443 4770 generic.go:334] "Generic (PLEG): container finished" podID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerID="4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" exitCode=0 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791505 4770 generic.go:334] "Generic (PLEG): container finished" podID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerID="4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" exitCode=0 Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791651 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerDied","Data":"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd"} Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791693 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerDied","Data":"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0"} Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791727 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b774e6dd-e8ac-4992-9129-a7fb80adff38","Type":"ContainerDied","Data":"7b9ff7fb8261e52bb8b502861ce5de31d4913954df25bd6b5de93b3b2e152647"} Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791749 4770 scope.go:117] "RemoveContainer" containerID="4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.791948 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.817040 4770 scope.go:117] "RemoveContainer" containerID="4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.848841 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.859244 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.861836 4770 scope.go:117] "RemoveContainer" containerID="4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.870542 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd\": container with ID starting with 4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd not found: ID does not exist" containerID="4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.870601 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd"} err="failed to get container status \"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd\": rpc error: code = NotFound desc = could not find container \"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd\": container with ID starting with 4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd not found: ID does not exist" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.870637 4770 scope.go:117] "RemoveContainer" containerID="4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.872826 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0\": container with ID starting with 4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0 not found: ID does not exist" containerID="4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.872864 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0"} err="failed to get container status \"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0\": rpc error: code = NotFound desc = could not find container \"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0\": container with ID starting with 4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0 not found: ID does not exist" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.872889 4770 scope.go:117] "RemoveContainer" containerID="4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.873766 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd"} err="failed to get container status \"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd\": rpc error: code = NotFound desc = could not find container \"4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd\": container with ID starting with 4f3e769024df8a3ea4ebbfaa528d6c2fef31944e7254688b8d902fa8dd4f60fd not found: ID does not exist" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.873847 4770 scope.go:117] "RemoveContainer" containerID="4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.874256 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0"} err="failed to get container status \"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0\": rpc error: code = NotFound desc = could not find container \"4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0\": container with ID starting with 4dec8de394bcc340fc38b8ad127a382c0e3d276b8ef59f36b0e46528fe9d62f0 not found: ID does not exist" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.874730 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.875214 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="init" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875235 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="init" Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.875303 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="probe" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875315 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="probe" Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.875379 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="dnsmasq-dns" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875389 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="dnsmasq-dns" Dec 09 11:55:27 crc kubenswrapper[4770]: E1209 11:55:27.875429 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="cinder-scheduler" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875438 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="cinder-scheduler" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875753 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="cinder-scheduler" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875782 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" containerName="probe" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.875799 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5124dab3-0ca0-40b9-beea-133de08b32bf" containerName="dnsmasq-dns" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.878383 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.884576 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 11:55:27 crc kubenswrapper[4770]: I1209 11:55:27.934618 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.059520 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b6jw\" (UniqueName: \"kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.059964 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.060056 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.060115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.060197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.060572 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163210 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163244 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163281 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163378 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163447 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b6jw\" (UniqueName: \"kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.163918 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.169582 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.171207 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.171319 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.177766 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.204540 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b6jw\" (UniqueName: \"kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw\") pod \"cinder-scheduler-0\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.240577 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.494822 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.500133 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.513596 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.683614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.684061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc6hh\" (UniqueName: \"kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.684243 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.786233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc6hh\" (UniqueName: \"kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.786319 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.786370 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.786852 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.787470 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.806320 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc6hh\" (UniqueName: \"kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh\") pod \"redhat-operators-c7f64\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.817401 4770 generic.go:334] "Generic (PLEG): container finished" podID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerID="8acac0c355671ed1bf61265f9f271acc9e87c858a0e6cd69df8c66b09d2912b3" exitCode=143 Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.817482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerDied","Data":"8acac0c355671ed1bf61265f9f271acc9e87c858a0e6cd69df8c66b09d2912b3"} Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.843081 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:28 crc kubenswrapper[4770]: I1209 11:55:28.887447 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.359008 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b774e6dd-e8ac-4992-9129-a7fb80adff38" path="/var/lib/kubelet/pods/b774e6dd-e8ac-4992-9129-a7fb80adff38/volumes" Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.421649 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.843644 4770 generic.go:334] "Generic (PLEG): container finished" podID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerID="4bd7d207b8c33342cc41b858d7d79c251e290cc4f55d34b6808f01af9542b265" exitCode=0 Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.843922 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerDied","Data":"4bd7d207b8c33342cc41b858d7d79c251e290cc4f55d34b6808f01af9542b265"} Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.843950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerStarted","Data":"d397bca54cecf8deb86a159900502c3a2ed35704b3b329a6da69dfc4cf79fe89"} Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.850110 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerStarted","Data":"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4"} Dec 09 11:55:29 crc kubenswrapper[4770]: I1209 11:55:29.850191 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerStarted","Data":"645ea8faf00828041c7d4a7dc07574facddded447ba73ced65e32861022a7189"} Dec 09 11:55:30 crc kubenswrapper[4770]: I1209 11:55:30.875453 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerStarted","Data":"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd"} Dec 09 11:55:30 crc kubenswrapper[4770]: I1209 11:55:30.890656 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6dcc946d68-b2gb6" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:51380->10.217.0.158:9311: read: connection reset by peer" Dec 09 11:55:30 crc kubenswrapper[4770]: I1209 11:55:30.891008 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6dcc946d68-b2gb6" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:51372->10.217.0.158:9311: read: connection reset by peer" Dec 09 11:55:30 crc kubenswrapper[4770]: I1209 11:55:30.892508 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerStarted","Data":"60db8b45019c1f065581edd251cfa6d44b7672ce7dbe914d94ee8c3bd50efc1a"} Dec 09 11:55:31 crc kubenswrapper[4770]: I1209 11:55:31.921836 4770 generic.go:334] "Generic (PLEG): container finished" podID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerID="b6ff8a331f1648332a1647a4235baa0db21ba20309d42c80141d918f7421332b" exitCode=0 Dec 09 11:55:31 crc kubenswrapper[4770]: I1209 11:55:31.922104 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerDied","Data":"b6ff8a331f1648332a1647a4235baa0db21ba20309d42c80141d918f7421332b"} Dec 09 11:55:31 crc kubenswrapper[4770]: I1209 11:55:31.927084 4770 generic.go:334] "Generic (PLEG): container finished" podID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerID="60db8b45019c1f065581edd251cfa6d44b7672ce7dbe914d94ee8c3bd50efc1a" exitCode=0 Dec 09 11:55:31 crc kubenswrapper[4770]: I1209 11:55:31.927291 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerDied","Data":"60db8b45019c1f065581edd251cfa6d44b7672ce7dbe914d94ee8c3bd50efc1a"} Dec 09 11:55:31 crc kubenswrapper[4770]: I1209 11:55:31.981499 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.981469618 podStartE2EDuration="4.981469618s" podCreationTimestamp="2025-12-09 11:55:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:31.972637454 +0000 UTC m=+1457.213395993" watchObservedRunningTime="2025-12-09 11:55:31.981469618 +0000 UTC m=+1457.222228137" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.631481 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.795795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8djt2\" (UniqueName: \"kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2\") pod \"8b273e04-19aa-4d8e-b82c-95f822c25677\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.796258 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data\") pod \"8b273e04-19aa-4d8e-b82c-95f822c25677\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.796352 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs\") pod \"8b273e04-19aa-4d8e-b82c-95f822c25677\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.796411 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle\") pod \"8b273e04-19aa-4d8e-b82c-95f822c25677\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.796432 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom\") pod \"8b273e04-19aa-4d8e-b82c-95f822c25677\" (UID: \"8b273e04-19aa-4d8e-b82c-95f822c25677\") " Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.797097 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs" (OuterVolumeSpecName: "logs") pod "8b273e04-19aa-4d8e-b82c-95f822c25677" (UID: "8b273e04-19aa-4d8e-b82c-95f822c25677"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.807583 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2" (OuterVolumeSpecName: "kube-api-access-8djt2") pod "8b273e04-19aa-4d8e-b82c-95f822c25677" (UID: "8b273e04-19aa-4d8e-b82c-95f822c25677"). InnerVolumeSpecName "kube-api-access-8djt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.816440 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8b273e04-19aa-4d8e-b82c-95f822c25677" (UID: "8b273e04-19aa-4d8e-b82c-95f822c25677"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.870849 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b273e04-19aa-4d8e-b82c-95f822c25677" (UID: "8b273e04-19aa-4d8e-b82c-95f822c25677"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.875709 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data" (OuterVolumeSpecName: "config-data") pod "8b273e04-19aa-4d8e-b82c-95f822c25677" (UID: "8b273e04-19aa-4d8e-b82c-95f822c25677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.897865 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.897919 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.897928 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8djt2\" (UniqueName: \"kubernetes.io/projected/8b273e04-19aa-4d8e-b82c-95f822c25677-kube-api-access-8djt2\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.897938 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b273e04-19aa-4d8e-b82c-95f822c25677-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.897947 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b273e04-19aa-4d8e-b82c-95f822c25677-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.944406 4770 generic.go:334] "Generic (PLEG): container finished" podID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerID="bd89441d5382fde55ff03f8f5a849f41ce9ccbe18c8191c1ab37523eab113b45" exitCode=0 Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.944501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerDied","Data":"bd89441d5382fde55ff03f8f5a849f41ce9ccbe18c8191c1ab37523eab113b45"} Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.961446 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcc946d68-b2gb6" event={"ID":"8b273e04-19aa-4d8e-b82c-95f822c25677","Type":"ContainerDied","Data":"d29ee20737c507128588a8d583c6927dbeb76a672fe68aaac20c80559b30faa2"} Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.961506 4770 scope.go:117] "RemoveContainer" containerID="b6ff8a331f1648332a1647a4235baa0db21ba20309d42c80141d918f7421332b" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.961644 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcc946d68-b2gb6" Dec 09 11:55:32 crc kubenswrapper[4770]: I1209 11:55:32.979240 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerStarted","Data":"cb296fa1e0f0134fb17f3b559f118a6838912bed0979b5119605a676ca2f5d7b"} Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.005646 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.014984 4770 scope.go:117] "RemoveContainer" containerID="8acac0c355671ed1bf61265f9f271acc9e87c858a0e6cd69df8c66b09d2912b3" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.022197 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6dcc946d68-b2gb6"] Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.030246 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c7f64" podStartSLOduration=2.41568133 podStartE2EDuration="5.030217988s" podCreationTimestamp="2025-12-09 11:55:28 +0000 UTC" firstStartedPulling="2025-12-09 11:55:29.845541123 +0000 UTC m=+1455.086299642" lastFinishedPulling="2025-12-09 11:55:32.460077781 +0000 UTC m=+1457.700836300" observedRunningTime="2025-12-09 11:55:33.014714984 +0000 UTC m=+1458.255473503" watchObservedRunningTime="2025-12-09 11:55:33.030217988 +0000 UTC m=+1458.270976507" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.241700 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.270134 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.306088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config\") pod \"1e501958-bf84-4a46-a0ee-526ebccdaf60\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.306152 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs\") pod \"1e501958-bf84-4a46-a0ee-526ebccdaf60\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.306195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk475\" (UniqueName: \"kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475\") pod \"1e501958-bf84-4a46-a0ee-526ebccdaf60\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.313117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1e501958-bf84-4a46-a0ee-526ebccdaf60" (UID: "1e501958-bf84-4a46-a0ee-526ebccdaf60"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.317245 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475" (OuterVolumeSpecName: "kube-api-access-dk475") pod "1e501958-bf84-4a46-a0ee-526ebccdaf60" (UID: "1e501958-bf84-4a46-a0ee-526ebccdaf60"). InnerVolumeSpecName "kube-api-access-dk475". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.359436 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" path="/var/lib/kubelet/pods/8b273e04-19aa-4d8e-b82c-95f822c25677/volumes" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.396164 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1e501958-bf84-4a46-a0ee-526ebccdaf60" (UID: "1e501958-bf84-4a46-a0ee-526ebccdaf60"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.408143 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle\") pod \"1e501958-bf84-4a46-a0ee-526ebccdaf60\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.408259 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config\") pod \"1e501958-bf84-4a46-a0ee-526ebccdaf60\" (UID: \"1e501958-bf84-4a46-a0ee-526ebccdaf60\") " Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.409037 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.409059 4770 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.409074 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk475\" (UniqueName: \"kubernetes.io/projected/1e501958-bf84-4a46-a0ee-526ebccdaf60-kube-api-access-dk475\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.624465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e501958-bf84-4a46-a0ee-526ebccdaf60" (UID: "1e501958-bf84-4a46-a0ee-526ebccdaf60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.627014 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.665770 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config" (OuterVolumeSpecName: "config") pod "1e501958-bf84-4a46-a0ee-526ebccdaf60" (UID: "1e501958-bf84-4a46-a0ee-526ebccdaf60"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.730039 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e501958-bf84-4a46-a0ee-526ebccdaf60-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.996869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8455957bd-72jgv" event={"ID":"1e501958-bf84-4a46-a0ee-526ebccdaf60","Type":"ContainerDied","Data":"f10b6f305df51099d0f045b5dc06277278500f84deb63a306cc51a46c77f2f2a"} Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.997333 4770 scope.go:117] "RemoveContainer" containerID="54e36573ebf61b52f64abf81a2d9c40f6e426c644c97d07e3601d928f6cb4c31" Dec 09 11:55:33 crc kubenswrapper[4770]: I1209 11:55:33.997572 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8455957bd-72jgv" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.018549 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.044332 4770 scope.go:117] "RemoveContainer" containerID="bd89441d5382fde55ff03f8f5a849f41ce9ccbe18c8191c1ab37523eab113b45" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.053270 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.097988 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8455957bd-72jgv"] Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.770354 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:55:34 crc kubenswrapper[4770]: E1209 11:55:34.771453 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-httpd" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771469 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-httpd" Dec 09 11:55:34 crc kubenswrapper[4770]: E1209 11:55:34.771489 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api-log" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771495 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api-log" Dec 09 11:55:34 crc kubenswrapper[4770]: E1209 11:55:34.771511 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771518 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api" Dec 09 11:55:34 crc kubenswrapper[4770]: E1209 11:55:34.771542 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-api" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771548 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-api" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771768 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-httpd" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771778 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" containerName="neutron-api" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771793 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.771812 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b273e04-19aa-4d8e-b82c-95f822c25677" containerName="barbican-api-log" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.774060 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.779940 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.780169 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.780336 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.811478 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.961806 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.961892 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzpcn\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.961957 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.962019 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.962054 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.962100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.962132 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:34 crc kubenswrapper[4770]: I1209 11:55:34.962171 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064214 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064272 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzpcn\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064303 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064323 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064350 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064387 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064408 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.064434 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.065128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.065394 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.070941 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.071181 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.071546 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.075300 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.078024 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.084837 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzpcn\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn\") pod \"swift-proxy-6b744fb85c-jt5h9\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.107433 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:35 crc kubenswrapper[4770]: I1209 11:55:35.357164 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e501958-bf84-4a46-a0ee-526ebccdaf60" path="/var/lib/kubelet/pods/1e501958-bf84-4a46-a0ee-526ebccdaf60/volumes" Dec 09 11:55:38 crc kubenswrapper[4770]: I1209 11:55:38.544428 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 11:55:38 crc kubenswrapper[4770]: I1209 11:55:38.843559 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:38 crc kubenswrapper[4770]: I1209 11:55:38.843621 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:38 crc kubenswrapper[4770]: I1209 11:55:38.897669 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:39 crc kubenswrapper[4770]: I1209 11:55:39.157492 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:39 crc kubenswrapper[4770]: I1209 11:55:39.202611 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:40 crc kubenswrapper[4770]: I1209 11:55:40.092672 4770 generic.go:334] "Generic (PLEG): container finished" podID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerID="fa1c1004130919ce9c120230332027edd32a27025efbe2c594c15d1d0a1c4eee" exitCode=137 Dec 09 11:55:40 crc kubenswrapper[4770]: I1209 11:55:40.094194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerDied","Data":"fa1c1004130919ce9c120230332027edd32a27025efbe2c594c15d1d0a1c4eee"} Dec 09 11:55:41 crc kubenswrapper[4770]: I1209 11:55:41.103236 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c7f64" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="registry-server" containerID="cri-o://cb296fa1e0f0134fb17f3b559f118a6838912bed0979b5119605a676ca2f5d7b" gracePeriod=2 Dec 09 11:55:43 crc kubenswrapper[4770]: I1209 11:55:43.125893 4770 generic.go:334] "Generic (PLEG): container finished" podID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerID="cb296fa1e0f0134fb17f3b559f118a6838912bed0979b5119605a676ca2f5d7b" exitCode=0 Dec 09 11:55:43 crc kubenswrapper[4770]: I1209 11:55:43.125978 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerDied","Data":"cb296fa1e0f0134fb17f3b559f118a6838912bed0979b5119605a676ca2f5d7b"} Dec 09 11:55:43 crc kubenswrapper[4770]: I1209 11:55:43.845077 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.143:3000/\": dial tcp 10.217.0.143:3000: connect: connection refused" Dec 09 11:55:46 crc kubenswrapper[4770]: I1209 11:55:46.769169 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.190735 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerStarted","Data":"01e26f55a994df9fdd65a6df86e3575b0a563a3c1822511ef020f6f8c9b66a18"} Dec 09 11:55:47 crc kubenswrapper[4770]: E1209 11:55:47.382118 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:2b4f8494513a3af102066fec5868ab167ac8664aceb2f0c639d7a0b60260a944" Dec 09 11:55:47 crc kubenswrapper[4770]: E1209 11:55:47.382322 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:2b4f8494513a3af102066fec5868ab167ac8664aceb2f0c639d7a0b60260a944,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54dh554hf8h59fh564h66ch5b5h5f8h575h54fhdbh4h665h584h685h689h56h64h567h5b4hd4hc4h595hdfh544hc5h544h56dh5cdh576h666h5cq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2nw4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(f6188401-2bb4-4ddc-a097-6b99f99df9e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:55:47 crc kubenswrapper[4770]: E1209 11:55:47.383534 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.709206 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.741831 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.909749 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.909848 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.909931 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities\") pod \"e9848bca-ca7a-405a-a5ee-72f829e6779a\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910029 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910079 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910096 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc6hh\" (UniqueName: \"kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh\") pod \"e9848bca-ca7a-405a-a5ee-72f829e6779a\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910194 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc2sk\" (UniqueName: \"kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910255 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content\") pod \"e9848bca-ca7a-405a-a5ee-72f829e6779a\" (UID: \"e9848bca-ca7a-405a-a5ee-72f829e6779a\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910293 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.910315 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd\") pod \"14c7c388-f457-44e4-ae7a-22904d6aa33c\" (UID: \"14c7c388-f457-44e4-ae7a-22904d6aa33c\") " Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.914717 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.915677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities" (OuterVolumeSpecName: "utilities") pod "e9848bca-ca7a-405a-a5ee-72f829e6779a" (UID: "e9848bca-ca7a-405a-a5ee-72f829e6779a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.915892 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.965152 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh" (OuterVolumeSpecName: "kube-api-access-rc6hh") pod "e9848bca-ca7a-405a-a5ee-72f829e6779a" (UID: "e9848bca-ca7a-405a-a5ee-72f829e6779a"). InnerVolumeSpecName "kube-api-access-rc6hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.965200 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts" (OuterVolumeSpecName: "scripts") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.965333 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk" (OuterVolumeSpecName: "kube-api-access-qc2sk") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "kube-api-access-qc2sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.975864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:47 crc kubenswrapper[4770]: I1209 11:55:47.998219 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012288 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9848bca-ca7a-405a-a5ee-72f829e6779a" (UID: "e9848bca-ca7a-405a-a5ee-72f829e6779a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012622 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc2sk\" (UniqueName: \"kubernetes.io/projected/14c7c388-f457-44e4-ae7a-22904d6aa33c-kube-api-access-qc2sk\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012651 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012661 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012670 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012679 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012690 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9848bca-ca7a-405a-a5ee-72f829e6779a-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012698 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c7c388-f457-44e4-ae7a-22904d6aa33c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012706 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.012716 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc6hh\" (UniqueName: \"kubernetes.io/projected/e9848bca-ca7a-405a-a5ee-72f829e6779a-kube-api-access-rc6hh\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.033140 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data" (OuterVolumeSpecName: "config-data") pod "14c7c388-f457-44e4-ae7a-22904d6aa33c" (UID: "14c7c388-f457-44e4-ae7a-22904d6aa33c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.116337 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c7c388-f457-44e4-ae7a-22904d6aa33c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.206437 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7f64" event={"ID":"e9848bca-ca7a-405a-a5ee-72f829e6779a","Type":"ContainerDied","Data":"d397bca54cecf8deb86a159900502c3a2ed35704b3b329a6da69dfc4cf79fe89"} Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.206491 4770 scope.go:117] "RemoveContainer" containerID="cb296fa1e0f0134fb17f3b559f118a6838912bed0979b5119605a676ca2f5d7b" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.206600 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7f64" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.210461 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerStarted","Data":"fe68553e529f03896be808d62dfc73b3485de5dc39de9017e810c7baf0d4521c"} Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.210503 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerStarted","Data":"b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90"} Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.211662 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.211696 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.233424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c7c388-f457-44e4-ae7a-22904d6aa33c","Type":"ContainerDied","Data":"e46aac983b5d1aa36d90b9a9b6934390c07a772ae9ec61632100586b57e7cae2"} Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.233653 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.263362 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6b744fb85c-jt5h9" podStartSLOduration=14.263336397 podStartE2EDuration="14.263336397s" podCreationTimestamp="2025-12-09 11:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:48.249320921 +0000 UTC m=+1473.490079450" watchObservedRunningTime="2025-12-09 11:55:48.263336397 +0000 UTC m=+1473.504094906" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.266456 4770 scope.go:117] "RemoveContainer" containerID="60db8b45019c1f065581edd251cfa6d44b7672ce7dbe914d94ee8c3bd50efc1a" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.266669 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:2b4f8494513a3af102066fec5868ab167ac8664aceb2f0c639d7a0b60260a944\\\"\"" pod="openstack/openstackclient" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.285710 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.296047 4770 scope.go:117] "RemoveContainer" containerID="4bd7d207b8c33342cc41b858d7d79c251e290cc4f55d34b6808f01af9542b265" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.302685 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c7f64"] Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.331515 4770 scope.go:117] "RemoveContainer" containerID="fa1c1004130919ce9c120230332027edd32a27025efbe2c594c15d1d0a1c4eee" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.352729 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.393041 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.407573 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408069 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="proxy-httpd" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408086 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="proxy-httpd" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408104 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="ceilometer-notification-agent" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408111 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="ceilometer-notification-agent" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408124 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="extract-content" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408131 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="extract-content" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408158 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="registry-server" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408166 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="registry-server" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408176 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="extract-utilities" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408182 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="extract-utilities" Dec 09 11:55:48 crc kubenswrapper[4770]: E1209 11:55:48.408193 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="sg-core" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408198 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="sg-core" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408414 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="ceilometer-notification-agent" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408433 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="sg-core" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408447 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" containerName="registry-server" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.408455 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" containerName="proxy-httpd" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.410441 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.413321 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.413438 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.429452 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.434417 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.434680 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.434807 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.434853 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjwc9\" (UniqueName: \"kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.434971 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.435046 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.435113 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.447994 4770 scope.go:117] "RemoveContainer" containerID="668171f0a1f1cde0fe4f4fe471f6859804937404f8461d50cc1a370f3b4c29dd" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.469388 4770 scope.go:117] "RemoveContainer" containerID="1e31cdf21f8cff9f7bd69dcb8a17c433f4dede72f5120c87e25f7f8705b55889" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.536532 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.536922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.536999 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537138 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537192 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjwc9\" (UniqueName: \"kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537505 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.537826 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.546335 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.546414 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.547021 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.550587 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.559262 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjwc9\" (UniqueName: \"kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9\") pod \"ceilometer-0\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " pod="openstack/ceilometer-0" Dec 09 11:55:48 crc kubenswrapper[4770]: I1209 11:55:48.746345 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.300022 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.360041 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c7c388-f457-44e4-ae7a-22904d6aa33c" path="/var/lib/kubelet/pods/14c7c388-f457-44e4-ae7a-22904d6aa33c/volumes" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.362588 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9848bca-ca7a-405a-a5ee-72f829e6779a" path="/var/lib/kubelet/pods/e9848bca-ca7a-405a-a5ee-72f829e6779a/volumes" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.642856 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-l8qd2"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.644738 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.651955 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-l8qd2"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.711424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km6j2\" (UniqueName: \"kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.711541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.759070 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-44v45"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.760525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.778753 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8c24-account-create-update-wxc89"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.780162 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.782922 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.788576 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8c24-account-create-update-wxc89"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.797470 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-44v45"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.818205 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4zvq\" (UniqueName: \"kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.819466 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.819559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hc9g\" (UniqueName: \"kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.819610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.819662 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.820159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km6j2\" (UniqueName: \"kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.820220 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.843833 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km6j2\" (UniqueName: \"kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2\") pod \"nova-api-db-create-l8qd2\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.922024 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hc9g\" (UniqueName: \"kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.922091 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.922127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.922317 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4zvq\" (UniqueName: \"kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.923608 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.923683 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.939705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4zvq\" (UniqueName: \"kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq\") pod \"nova-cell0-db-create-44v45\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.951870 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hc9g\" (UniqueName: \"kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g\") pod \"nova-api-8c24-account-create-update-wxc89\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.963298 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xhvz6"] Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.965097 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.967409 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:49 crc kubenswrapper[4770]: I1209 11:55:49.990216 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xhvz6"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.005715 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-17c1-account-create-update-vnk7f"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.010685 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.030166 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.052193 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-17c1-account-create-update-vnk7f"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.125396 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.141613 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.141677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.141833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm4d5\" (UniqueName: \"kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.145346 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v4wj\" (UniqueName: \"kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.168825 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7adb-account-create-update-nkbct"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.170553 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.173968 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.181057 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7adb-account-create-update-nkbct"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.227954 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.246683 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5px8n\" (UniqueName: \"kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.246932 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.246972 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.247042 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.247184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm4d5\" (UniqueName: \"kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.247271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v4wj\" (UniqueName: \"kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.248765 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.251169 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.280552 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v4wj\" (UniqueName: \"kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj\") pod \"nova-cell0-17c1-account-create-update-vnk7f\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.289580 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm4d5\" (UniqueName: \"kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5\") pod \"nova-cell1-db-create-xhvz6\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.293051 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerStarted","Data":"af8f2cb0c684d4a22cfce2984e7a27e5e53dab1913cc27351bef3652477fae17"} Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.293145 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerStarted","Data":"ced6a426d1230370028e4de570483002be32e6a2bca00d23176595cd237e62a0"} Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.348661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5px8n\" (UniqueName: \"kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.348890 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.351501 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.353596 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.364449 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.369083 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5px8n\" (UniqueName: \"kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n\") pod \"nova-cell1-7adb-account-create-update-nkbct\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.494349 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.608939 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-l8qd2"] Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.725595 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-44v45"] Dec 09 11:55:50 crc kubenswrapper[4770]: W1209 11:55:50.740658 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1672e55_70d6_4e2e_9900_c2744265763b.slice/crio-eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159 WatchSource:0}: Error finding container eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159: Status 404 returned error can't find the container with id eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159 Dec 09 11:55:50 crc kubenswrapper[4770]: I1209 11:55:50.935444 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8c24-account-create-update-wxc89"] Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.082764 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xhvz6"] Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.138296 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-17c1-account-create-update-vnk7f"] Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.305454 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8c24-account-create-update-wxc89" event={"ID":"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9","Type":"ContainerStarted","Data":"2bbe5d78041ffbdd917a78be6b066422f3130937304df0fb5318bb5b4e09ea2f"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.306002 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8c24-account-create-update-wxc89" event={"ID":"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9","Type":"ContainerStarted","Data":"4ae790887a4ef6585cf9832adc04ca4dcea04a38fc2df6486900497bf014ff95"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.308281 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerStarted","Data":"89f03724fff6c534e61057eb4e541ada12463fbab10e60ff30ca842ba4a9c22c"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.314568 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7adb-account-create-update-nkbct"] Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.315170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l8qd2" event={"ID":"1e46e9fb-3610-414e-acef-28323578d294","Type":"ContainerStarted","Data":"fa7bcf45aabcd3aa07a85a0fa2eb580f33813d2af8b453470cd5519a7b3285cd"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.315228 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l8qd2" event={"ID":"1e46e9fb-3610-414e-acef-28323578d294","Type":"ContainerStarted","Data":"a473752466b82b29c0111d394eac691283a27376fcc0cd2d7b18f70e123eee6f"} Dec 09 11:55:51 crc kubenswrapper[4770]: W1209 11:55:51.319606 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3db1fbc4_24a1_4662_b146_064d33c9d5d6.slice/crio-813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b WatchSource:0}: Error finding container 813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b: Status 404 returned error can't find the container with id 813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.325721 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-44v45" event={"ID":"a1672e55-70d6-4e2e-9900-c2744265763b","Type":"ContainerStarted","Data":"baec3dc745b1f2b4180b68e8ccb0a514d518f4ffcd36092950684e12e387a258"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.325791 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-44v45" event={"ID":"a1672e55-70d6-4e2e-9900-c2744265763b","Type":"ContainerStarted","Data":"eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.333633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" event={"ID":"2c47befa-1ab8-4d48-b43f-806e2335066a","Type":"ContainerStarted","Data":"c8fd697ef9bb553a8cf3ba1e224712113852589bcde75b1a1b27bffcddf8ff75"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.356970 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-44v45" podStartSLOduration=2.356946329 podStartE2EDuration="2.356946329s" podCreationTimestamp="2025-12-09 11:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:51.34988648 +0000 UTC m=+1476.590645009" watchObservedRunningTime="2025-12-09 11:55:51.356946329 +0000 UTC m=+1476.597704848" Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.357514 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xhvz6" event={"ID":"508d8422-6ead-4541-b5e3-b16337546e7d","Type":"ContainerStarted","Data":"97784227d19d232a70e13d223f833bfa3697d768316466c286119becb5239a78"} Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.361093 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8c24-account-create-update-wxc89" podStartSLOduration=2.361068724 podStartE2EDuration="2.361068724s" podCreationTimestamp="2025-12-09 11:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:51.331503603 +0000 UTC m=+1476.572262132" watchObservedRunningTime="2025-12-09 11:55:51.361068724 +0000 UTC m=+1476.601827243" Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.389792 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-l8qd2" podStartSLOduration=2.389767053 podStartE2EDuration="2.389767053s" podCreationTimestamp="2025-12-09 11:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:51.373568851 +0000 UTC m=+1476.614327370" watchObservedRunningTime="2025-12-09 11:55:51.389767053 +0000 UTC m=+1476.630525572" Dec 09 11:55:51 crc kubenswrapper[4770]: I1209 11:55:51.891122 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.361619 4770 generic.go:334] "Generic (PLEG): container finished" podID="a1672e55-70d6-4e2e-9900-c2744265763b" containerID="baec3dc745b1f2b4180b68e8ccb0a514d518f4ffcd36092950684e12e387a258" exitCode=0 Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.361722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-44v45" event={"ID":"a1672e55-70d6-4e2e-9900-c2744265763b","Type":"ContainerDied","Data":"baec3dc745b1f2b4180b68e8ccb0a514d518f4ffcd36092950684e12e387a258"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.373106 4770 generic.go:334] "Generic (PLEG): container finished" podID="2c47befa-1ab8-4d48-b43f-806e2335066a" containerID="b6dee3f17438b10b86b6871f88dfe7765fc45e68f4da390d513c0bd71565c8c6" exitCode=0 Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.373213 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" event={"ID":"2c47befa-1ab8-4d48-b43f-806e2335066a","Type":"ContainerDied","Data":"b6dee3f17438b10b86b6871f88dfe7765fc45e68f4da390d513c0bd71565c8c6"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.375829 4770 generic.go:334] "Generic (PLEG): container finished" podID="508d8422-6ead-4541-b5e3-b16337546e7d" containerID="b2d9d48cbe19f172b7ac2c4f7e5d87d39111563cf778bcbc8050188702d7c75f" exitCode=0 Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.375912 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xhvz6" event={"ID":"508d8422-6ead-4541-b5e3-b16337546e7d","Type":"ContainerDied","Data":"b2d9d48cbe19f172b7ac2c4f7e5d87d39111563cf778bcbc8050188702d7c75f"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.380894 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" event={"ID":"3db1fbc4-24a1-4662-b146-064d33c9d5d6","Type":"ContainerStarted","Data":"07184d4ac9451f9079837acb457f303a341323b0ecb029c30cb51673a1719aae"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.380993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" event={"ID":"3db1fbc4-24a1-4662-b146-064d33c9d5d6","Type":"ContainerStarted","Data":"813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.385187 4770 generic.go:334] "Generic (PLEG): container finished" podID="a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" containerID="2bbe5d78041ffbdd917a78be6b066422f3130937304df0fb5318bb5b4e09ea2f" exitCode=0 Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.385288 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8c24-account-create-update-wxc89" event={"ID":"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9","Type":"ContainerDied","Data":"2bbe5d78041ffbdd917a78be6b066422f3130937304df0fb5318bb5b4e09ea2f"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.395823 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerStarted","Data":"f55f463367ab1c8bc35972b6a41d9cb825b48d3909ce0451765716ca6c0a0267"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.398293 4770 generic.go:334] "Generic (PLEG): container finished" podID="1e46e9fb-3610-414e-acef-28323578d294" containerID="fa7bcf45aabcd3aa07a85a0fa2eb580f33813d2af8b453470cd5519a7b3285cd" exitCode=0 Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.398376 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l8qd2" event={"ID":"1e46e9fb-3610-414e-acef-28323578d294","Type":"ContainerDied","Data":"fa7bcf45aabcd3aa07a85a0fa2eb580f33813d2af8b453470cd5519a7b3285cd"} Dec 09 11:55:52 crc kubenswrapper[4770]: I1209 11:55:52.464725 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" podStartSLOduration=2.464702897 podStartE2EDuration="2.464702897s" podCreationTimestamp="2025-12-09 11:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:55:52.453354879 +0000 UTC m=+1477.694113398" watchObservedRunningTime="2025-12-09 11:55:52.464702897 +0000 UTC m=+1477.705461416" Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.417819 4770 generic.go:334] "Generic (PLEG): container finished" podID="3db1fbc4-24a1-4662-b146-064d33c9d5d6" containerID="07184d4ac9451f9079837acb457f303a341323b0ecb029c30cb51673a1719aae" exitCode=0 Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.417941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" event={"ID":"3db1fbc4-24a1-4662-b146-064d33c9d5d6","Type":"ContainerDied","Data":"07184d4ac9451f9079837acb457f303a341323b0ecb029c30cb51673a1719aae"} Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.423498 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-central-agent" containerID="cri-o://af8f2cb0c684d4a22cfce2984e7a27e5e53dab1913cc27351bef3652477fae17" gracePeriod=30 Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.423956 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="proxy-httpd" containerID="cri-o://0ca6ad8ec1507c0162d39456fba3bac7cf36c9d4e3c5e8ad7d0894c5c5cc16b6" gracePeriod=30 Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.424090 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-notification-agent" containerID="cri-o://89f03724fff6c534e61057eb4e541ada12463fbab10e60ff30ca842ba4a9c22c" gracePeriod=30 Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.424137 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="sg-core" containerID="cri-o://f55f463367ab1c8bc35972b6a41d9cb825b48d3909ce0451765716ca6c0a0267" gracePeriod=30 Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.485589 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6447883939999999 podStartE2EDuration="5.485563019s" podCreationTimestamp="2025-12-09 11:55:48 +0000 UTC" firstStartedPulling="2025-12-09 11:55:49.304720219 +0000 UTC m=+1474.545478738" lastFinishedPulling="2025-12-09 11:55:53.145494844 +0000 UTC m=+1478.386253363" observedRunningTime="2025-12-09 11:55:53.462275087 +0000 UTC m=+1478.703033606" watchObservedRunningTime="2025-12-09 11:55:53.485563019 +0000 UTC m=+1478.726321538" Dec 09 11:55:53 crc kubenswrapper[4770]: I1209 11:55:53.877984 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.012093 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.028189 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm4d5\" (UniqueName: \"kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5\") pod \"508d8422-6ead-4541-b5e3-b16337546e7d\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.028399 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts\") pod \"508d8422-6ead-4541-b5e3-b16337546e7d\" (UID: \"508d8422-6ead-4541-b5e3-b16337546e7d\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.032248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "508d8422-6ead-4541-b5e3-b16337546e7d" (UID: "508d8422-6ead-4541-b5e3-b16337546e7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.033422 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.035929 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.036562 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5" (OuterVolumeSpecName: "kube-api-access-pm4d5") pod "508d8422-6ead-4541-b5e3-b16337546e7d" (UID: "508d8422-6ead-4541-b5e3-b16337546e7d"). InnerVolumeSpecName "kube-api-access-pm4d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.056008 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm4d5\" (UniqueName: \"kubernetes.io/projected/508d8422-6ead-4541-b5e3-b16337546e7d-kube-api-access-pm4d5\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.056141 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508d8422-6ead-4541-b5e3-b16337546e7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.068433 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.168771 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hc9g\" (UniqueName: \"kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g\") pod \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169197 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts\") pod \"1e46e9fb-3610-414e-acef-28323578d294\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169343 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts\") pod \"2c47befa-1ab8-4d48-b43f-806e2335066a\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169573 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v4wj\" (UniqueName: \"kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj\") pod \"2c47befa-1ab8-4d48-b43f-806e2335066a\" (UID: \"2c47befa-1ab8-4d48-b43f-806e2335066a\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km6j2\" (UniqueName: \"kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2\") pod \"1e46e9fb-3610-414e-acef-28323578d294\" (UID: \"1e46e9fb-3610-414e-acef-28323578d294\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169680 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts\") pod \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\" (UID: \"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.169967 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e46e9fb-3610-414e-acef-28323578d294" (UID: "1e46e9fb-3610-414e-acef-28323578d294"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.170385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" (UID: "a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.170485 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e46e9fb-3610-414e-acef-28323578d294-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.171276 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c47befa-1ab8-4d48-b43f-806e2335066a" (UID: "2c47befa-1ab8-4d48-b43f-806e2335066a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.180486 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2" (OuterVolumeSpecName: "kube-api-access-km6j2") pod "1e46e9fb-3610-414e-acef-28323578d294" (UID: "1e46e9fb-3610-414e-acef-28323578d294"). InnerVolumeSpecName "kube-api-access-km6j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.184985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g" (OuterVolumeSpecName: "kube-api-access-5hc9g") pod "a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" (UID: "a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9"). InnerVolumeSpecName "kube-api-access-5hc9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.196225 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj" (OuterVolumeSpecName: "kube-api-access-8v4wj") pod "2c47befa-1ab8-4d48-b43f-806e2335066a" (UID: "2c47befa-1ab8-4d48-b43f-806e2335066a"). InnerVolumeSpecName "kube-api-access-8v4wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.271267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zvq\" (UniqueName: \"kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq\") pod \"a1672e55-70d6-4e2e-9900-c2744265763b\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.271639 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts\") pod \"a1672e55-70d6-4e2e-9900-c2744265763b\" (UID: \"a1672e55-70d6-4e2e-9900-c2744265763b\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272056 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272070 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hc9g\" (UniqueName: \"kubernetes.io/projected/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9-kube-api-access-5hc9g\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272083 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c47befa-1ab8-4d48-b43f-806e2335066a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272091 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v4wj\" (UniqueName: \"kubernetes.io/projected/2c47befa-1ab8-4d48-b43f-806e2335066a-kube-api-access-8v4wj\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272100 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km6j2\" (UniqueName: \"kubernetes.io/projected/1e46e9fb-3610-414e-acef-28323578d294-kube-api-access-km6j2\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.272279 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1672e55-70d6-4e2e-9900-c2744265763b" (UID: "a1672e55-70d6-4e2e-9900-c2744265763b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.277823 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq" (OuterVolumeSpecName: "kube-api-access-x4zvq") pod "a1672e55-70d6-4e2e-9900-c2744265763b" (UID: "a1672e55-70d6-4e2e-9900-c2744265763b"). InnerVolumeSpecName "kube-api-access-x4zvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.374243 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1672e55-70d6-4e2e-9900-c2744265763b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.374294 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zvq\" (UniqueName: \"kubernetes.io/projected/a1672e55-70d6-4e2e-9900-c2744265763b-kube-api-access-x4zvq\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.440217 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-l8qd2" event={"ID":"1e46e9fb-3610-414e-acef-28323578d294","Type":"ContainerDied","Data":"a473752466b82b29c0111d394eac691283a27376fcc0cd2d7b18f70e123eee6f"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.440596 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a473752466b82b29c0111d394eac691283a27376fcc0cd2d7b18f70e123eee6f" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.440671 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-l8qd2" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.444538 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-44v45" event={"ID":"a1672e55-70d6-4e2e-9900-c2744265763b","Type":"ContainerDied","Data":"eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.444617 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaa61847da509320e1a0c91b147938d86538e11c0736b9c9ad80cb384074f159" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.444595 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-44v45" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.448501 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.448543 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-17c1-account-create-update-vnk7f" event={"ID":"2c47befa-1ab8-4d48-b43f-806e2335066a","Type":"ContainerDied","Data":"c8fd697ef9bb553a8cf3ba1e224712113852589bcde75b1a1b27bffcddf8ff75"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.448595 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8fd697ef9bb553a8cf3ba1e224712113852589bcde75b1a1b27bffcddf8ff75" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.454978 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xhvz6" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.455035 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xhvz6" event={"ID":"508d8422-6ead-4541-b5e3-b16337546e7d","Type":"ContainerDied","Data":"97784227d19d232a70e13d223f833bfa3697d768316466c286119becb5239a78"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.455270 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97784227d19d232a70e13d223f833bfa3697d768316466c286119becb5239a78" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.456799 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8c24-account-create-update-wxc89" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.456883 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8c24-account-create-update-wxc89" event={"ID":"a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9","Type":"ContainerDied","Data":"4ae790887a4ef6585cf9832adc04ca4dcea04a38fc2df6486900497bf014ff95"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.456960 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ae790887a4ef6585cf9832adc04ca4dcea04a38fc2df6486900497bf014ff95" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.461207 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerID="f55f463367ab1c8bc35972b6a41d9cb825b48d3909ce0451765716ca6c0a0267" exitCode=2 Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.461246 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerStarted","Data":"0ca6ad8ec1507c0162d39456fba3bac7cf36c9d4e3c5e8ad7d0894c5c5cc16b6"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.461315 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerDied","Data":"f55f463367ab1c8bc35972b6a41d9cb825b48d3909ce0451765716ca6c0a0267"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.461263 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerID="89f03724fff6c534e61057eb4e541ada12463fbab10e60ff30ca842ba4a9c22c" exitCode=0 Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.461338 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerDied","Data":"89f03724fff6c534e61057eb4e541ada12463fbab10e60ff30ca842ba4a9c22c"} Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.800874 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.984733 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5px8n\" (UniqueName: \"kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n\") pod \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.985130 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts\") pod \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\" (UID: \"3db1fbc4-24a1-4662-b146-064d33c9d5d6\") " Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.985710 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3db1fbc4-24a1-4662-b146-064d33c9d5d6" (UID: "3db1fbc4-24a1-4662-b146-064d33c9d5d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:55:54 crc kubenswrapper[4770]: I1209 11:55:54.990735 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n" (OuterVolumeSpecName: "kube-api-access-5px8n") pod "3db1fbc4-24a1-4662-b146-064d33c9d5d6" (UID: "3db1fbc4-24a1-4662-b146-064d33c9d5d6"). InnerVolumeSpecName "kube-api-access-5px8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.087633 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5px8n\" (UniqueName: \"kubernetes.io/projected/3db1fbc4-24a1-4662-b146-064d33c9d5d6-kube-api-access-5px8n\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.087669 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db1fbc4-24a1-4662-b146-064d33c9d5d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.116203 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.118490 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.471107 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.471108 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7adb-account-create-update-nkbct" event={"ID":"3db1fbc4-24a1-4662-b146-064d33c9d5d6","Type":"ContainerDied","Data":"813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b"} Dec 09 11:55:55 crc kubenswrapper[4770]: I1209 11:55:55.471163 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="813f81f106e7871ba36b1aa917f23796c8acaa9785abe704d0c35ecaf3a1330b" Dec 09 11:55:56 crc kubenswrapper[4770]: I1209 11:55:56.484958 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerID="af8f2cb0c684d4a22cfce2984e7a27e5e53dab1913cc27351bef3652477fae17" exitCode=0 Dec 09 11:55:56 crc kubenswrapper[4770]: I1209 11:55:56.484971 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerDied","Data":"af8f2cb0c684d4a22cfce2984e7a27e5e53dab1913cc27351bef3652477fae17"} Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.202694 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8twmk"] Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.203678 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d8422-6ead-4541-b5e3-b16337546e7d" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.203693 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d8422-6ead-4541-b5e3-b16337546e7d" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.203713 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1672e55-70d6-4e2e-9900-c2744265763b" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.203719 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1672e55-70d6-4e2e-9900-c2744265763b" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.203728 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3db1fbc4-24a1-4662-b146-064d33c9d5d6" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.203734 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3db1fbc4-24a1-4662-b146-064d33c9d5d6" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.203747 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c47befa-1ab8-4d48-b43f-806e2335066a" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.203753 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c47befa-1ab8-4d48-b43f-806e2335066a" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.203780 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.205994 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: E1209 11:56:00.206111 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e46e9fb-3610-414e-acef-28323578d294" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.206126 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e46e9fb-3610-414e-acef-28323578d294" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207590 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c47befa-1ab8-4d48-b43f-806e2335066a" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207636 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3db1fbc4-24a1-4662-b146-064d33c9d5d6" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207645 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1672e55-70d6-4e2e-9900-c2744265763b" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207663 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e46e9fb-3610-414e-acef-28323578d294" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207706 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="508d8422-6ead-4541-b5e3-b16337546e7d" containerName="mariadb-database-create" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.207732 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" containerName="mariadb-account-create-update" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.209035 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.220746 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.220806 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gdt78" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.222854 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.238837 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8twmk"] Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.318752 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.318843 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blpgv\" (UniqueName: \"kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.318970 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.319336 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.421687 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.422028 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.422124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blpgv\" (UniqueName: \"kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.422228 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.427488 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.428274 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.429946 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.453760 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blpgv\" (UniqueName: \"kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv\") pod \"nova-cell0-conductor-db-sync-8twmk\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:00 crc kubenswrapper[4770]: I1209 11:56:00.557981 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:01 crc kubenswrapper[4770]: I1209 11:56:01.076402 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8twmk"] Dec 09 11:56:01 crc kubenswrapper[4770]: I1209 11:56:01.534486 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8twmk" event={"ID":"eb132594-6dbe-452f-ab35-8337509cd95d","Type":"ContainerStarted","Data":"796c18a5311419f84f129ca3ce4aba68e8a6be7c1c17d1f6ddf11958587ebf38"} Dec 09 11:56:04 crc kubenswrapper[4770]: I1209 11:56:04.594633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f6188401-2bb4-4ddc-a097-6b99f99df9e8","Type":"ContainerStarted","Data":"faf973a0c6d5bcf16a241dc80a13699d03f0a874da0549eb1f5e4001556d13ae"} Dec 09 11:56:04 crc kubenswrapper[4770]: I1209 11:56:04.626085 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.6134578939999997 podStartE2EDuration="40.626063867s" podCreationTimestamp="2025-12-09 11:55:24 +0000 UTC" firstStartedPulling="2025-12-09 11:55:25.847180367 +0000 UTC m=+1451.087938886" lastFinishedPulling="2025-12-09 11:56:03.85978634 +0000 UTC m=+1489.100544859" observedRunningTime="2025-12-09 11:56:04.617452389 +0000 UTC m=+1489.858210908" watchObservedRunningTime="2025-12-09 11:56:04.626063867 +0000 UTC m=+1489.866822386" Dec 09 11:56:06 crc kubenswrapper[4770]: I1209 11:56:06.281612 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:06 crc kubenswrapper[4770]: I1209 11:56:06.282458 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-log" containerID="cri-o://7f7eaccb9a37b54a7a21a7a819223a1106f81f476a8f56d4926ef19ebddbb708" gracePeriod=30 Dec 09 11:56:06 crc kubenswrapper[4770]: I1209 11:56:06.282555 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-httpd" containerID="cri-o://334ccf338ffd966eccc95375cc7b064fdbb4ecd1a1b5f283b0cb6961e3e4d9aa" gracePeriod=30 Dec 09 11:56:06 crc kubenswrapper[4770]: I1209 11:56:06.626803 4770 generic.go:334] "Generic (PLEG): container finished" podID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerID="7f7eaccb9a37b54a7a21a7a819223a1106f81f476a8f56d4926ef19ebddbb708" exitCode=143 Dec 09 11:56:06 crc kubenswrapper[4770]: I1209 11:56:06.626862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerDied","Data":"7f7eaccb9a37b54a7a21a7a819223a1106f81f476a8f56d4926ef19ebddbb708"} Dec 09 11:56:07 crc kubenswrapper[4770]: I1209 11:56:07.534494 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:07 crc kubenswrapper[4770]: I1209 11:56:07.534828 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-log" containerID="cri-o://b0b0312f945dfc048ae1a1bc134866f989faa293193ee9b494cd073af664a917" gracePeriod=30 Dec 09 11:56:07 crc kubenswrapper[4770]: I1209 11:56:07.534852 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-httpd" containerID="cri-o://34e7162111d82c416b36d7cac62876c550b4c2a95c5dd62c86e63169addfa649" gracePeriod=30 Dec 09 11:56:08 crc kubenswrapper[4770]: I1209 11:56:08.655491 4770 generic.go:334] "Generic (PLEG): container finished" podID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerID="b0b0312f945dfc048ae1a1bc134866f989faa293193ee9b494cd073af664a917" exitCode=143 Dec 09 11:56:08 crc kubenswrapper[4770]: I1209 11:56:08.655674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerDied","Data":"b0b0312f945dfc048ae1a1bc134866f989faa293193ee9b494cd073af664a917"} Dec 09 11:56:09 crc kubenswrapper[4770]: I1209 11:56:09.671501 4770 generic.go:334] "Generic (PLEG): container finished" podID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerID="334ccf338ffd966eccc95375cc7b064fdbb4ecd1a1b5f283b0cb6961e3e4d9aa" exitCode=0 Dec 09 11:56:09 crc kubenswrapper[4770]: I1209 11:56:09.671577 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerDied","Data":"334ccf338ffd966eccc95375cc7b064fdbb4ecd1a1b5f283b0cb6961e3e4d9aa"} Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.230298 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": dial tcp 10.217.0.147:9292: connect: connection refused" Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.230311 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": dial tcp 10.217.0.147:9292: connect: connection refused" Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.612079 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": dial tcp 10.217.0.148:9292: connect: connection refused" Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.612209 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": dial tcp 10.217.0.148:9292: connect: connection refused" Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.699330 4770 generic.go:334] "Generic (PLEG): container finished" podID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerID="34e7162111d82c416b36d7cac62876c550b4c2a95c5dd62c86e63169addfa649" exitCode=0 Dec 09 11:56:11 crc kubenswrapper[4770]: I1209 11:56:11.699380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerDied","Data":"34e7162111d82c416b36d7cac62876c550b4c2a95c5dd62c86e63169addfa649"} Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.713447 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8twmk" event={"ID":"eb132594-6dbe-452f-ab35-8337509cd95d","Type":"ContainerStarted","Data":"f483d04a4a2d6845118ac8581f5705bae9286fde5e96eb516d573219ca6e36da"} Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.740562 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8twmk" podStartSLOduration=1.391025214 podStartE2EDuration="12.74054239s" podCreationTimestamp="2025-12-09 11:56:00 +0000 UTC" firstStartedPulling="2025-12-09 11:56:01.072825984 +0000 UTC m=+1486.313584503" lastFinishedPulling="2025-12-09 11:56:12.42234316 +0000 UTC m=+1497.663101679" observedRunningTime="2025-12-09 11:56:12.737268917 +0000 UTC m=+1497.978027466" watchObservedRunningTime="2025-12-09 11:56:12.74054239 +0000 UTC m=+1497.981300919" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.766627 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.906649 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.906768 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.906804 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.906999 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.907030 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.907068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2xzd\" (UniqueName: \"kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.907109 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.907198 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs\") pod \"81346ca2-bec1-4d0c-900a-2af788284bc7\" (UID: \"81346ca2-bec1-4d0c-900a-2af788284bc7\") " Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.910356 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.910651 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs" (OuterVolumeSpecName: "logs") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.926601 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.931858 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd" (OuterVolumeSpecName: "kube-api-access-w2xzd") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "kube-api-access-w2xzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:56:12 crc kubenswrapper[4770]: I1209 11:56:12.940770 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts" (OuterVolumeSpecName: "scripts") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.013610 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2xzd\" (UniqueName: \"kubernetes.io/projected/81346ca2-bec1-4d0c-900a-2af788284bc7-kube-api-access-w2xzd\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.013668 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.013713 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.013728 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.013740 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81346ca2-bec1-4d0c-900a-2af788284bc7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.048757 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.074955 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.112727 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.116491 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.116538 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.170037 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data" (OuterVolumeSpecName: "config-data") pod "81346ca2-bec1-4d0c-900a-2af788284bc7" (UID: "81346ca2-bec1-4d0c-900a-2af788284bc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.177486 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.217524 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.217623 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.217664 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.217798 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.217963 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218039 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7gpd\" (UniqueName: \"kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218082 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218108 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle\") pod \"4043b25c-3d5f-4f4d-8eea-9990caab521c\" (UID: \"4043b25c-3d5f-4f4d-8eea-9990caab521c\") " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218643 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218668 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81346ca2-bec1-4d0c-900a-2af788284bc7-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.218961 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs" (OuterVolumeSpecName: "logs") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.225059 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.225364 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.239530 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts" (OuterVolumeSpecName: "scripts") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.247172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd" (OuterVolumeSpecName: "kube-api-access-b7gpd") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "kube-api-access-b7gpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.281756 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data" (OuterVolumeSpecName: "config-data") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.306416 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.309787 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4043b25c-3d5f-4f4d-8eea-9990caab521c" (UID: "4043b25c-3d5f-4f4d-8eea-9990caab521c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320875 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320936 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7gpd\" (UniqueName: \"kubernetes.io/projected/4043b25c-3d5f-4f4d-8eea-9990caab521c-kube-api-access-b7gpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320950 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320964 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320974 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320981 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320990 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4043b25c-3d5f-4f4d-8eea-9990caab521c-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.320997 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4043b25c-3d5f-4f4d-8eea-9990caab521c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.342339 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.423180 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.724804 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81346ca2-bec1-4d0c-900a-2af788284bc7","Type":"ContainerDied","Data":"fedd2f06e5ed46971a39c6dc0e03d77a5424ea16cebd322156baaf4adf7555c3"} Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.724880 4770 scope.go:117] "RemoveContainer" containerID="34e7162111d82c416b36d7cac62876c550b4c2a95c5dd62c86e63169addfa649" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.725111 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.730087 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4043b25c-3d5f-4f4d-8eea-9990caab521c","Type":"ContainerDied","Data":"9014350d818bc33f317c92132e8085363c6a71a504debb4b2d6e1454db4996bc"} Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.730118 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.759654 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.775606 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.781887 4770 scope.go:117] "RemoveContainer" containerID="b0b0312f945dfc048ae1a1bc134866f989faa293193ee9b494cd073af664a917" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.800830 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.812028 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.829630 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: E1209 11:56:13.830632 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.830757 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: E1209 11:56:13.830856 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831004 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: E1209 11:56:13.831102 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831174 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: E1209 11:56:13.831265 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831342 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831693 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831814 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-log" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.831912 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.832004 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" containerName="glance-httpd" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.833671 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.856252 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fvqzp" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.856496 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.856657 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.856812 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.866757 4770 scope.go:117] "RemoveContainer" containerID="334ccf338ffd966eccc95375cc7b064fdbb4ecd1a1b5f283b0cb6961e3e4d9aa" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.875240 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.909638 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.912435 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.915891 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.916238 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937435 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937506 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937533 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937561 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937580 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z92n\" (UniqueName: \"kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937611 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937630 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.937705 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.941960 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:13 crc kubenswrapper[4770]: I1209 11:56:13.955468 4770 scope.go:117] "RemoveContainer" containerID="7f7eaccb9a37b54a7a21a7a819223a1106f81f476a8f56d4926ef19ebddbb708" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.039647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040127 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq4dr\" (UniqueName: \"kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040183 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040313 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040380 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040443 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040634 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040699 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040735 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040762 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040792 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040875 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040919 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z92n\" (UniqueName: \"kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.040985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.041008 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.041161 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.041566 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.041665 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.046660 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.047257 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.055223 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.056632 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.062077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z92n\" (UniqueName: \"kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.078137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143055 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143131 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143404 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq4dr\" (UniqueName: \"kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143530 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.143577 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.144044 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.144044 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.144149 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.150173 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.150878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.151254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.151353 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.163854 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq4dr\" (UniqueName: \"kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.181627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.197382 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.255244 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:56:14 crc kubenswrapper[4770]: I1209 11:56:14.811744 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:56:15 crc kubenswrapper[4770]: I1209 11:56:15.043274 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:56:15 crc kubenswrapper[4770]: I1209 11:56:15.361525 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4043b25c-3d5f-4f4d-8eea-9990caab521c" path="/var/lib/kubelet/pods/4043b25c-3d5f-4f4d-8eea-9990caab521c/volumes" Dec 09 11:56:15 crc kubenswrapper[4770]: I1209 11:56:15.362962 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81346ca2-bec1-4d0c-900a-2af788284bc7" path="/var/lib/kubelet/pods/81346ca2-bec1-4d0c-900a-2af788284bc7/volumes" Dec 09 11:56:15 crc kubenswrapper[4770]: I1209 11:56:15.756389 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerStarted","Data":"a52884cbb85f13029b50a0ff7af99099b7aabcafb4a808f4b3590340e449e81b"} Dec 09 11:56:15 crc kubenswrapper[4770]: I1209 11:56:15.758682 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerStarted","Data":"e663a780d9b06b4e39849bb7b0c4661407033e945777e41c7fb8eca7c8e320f3"} Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.771185 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerStarted","Data":"eebed45a512ca372586490859e850a999b9d581d64647896a65dfb070e1af5fb"} Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.773336 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerStarted","Data":"7569e5bc00921bcf0e4cb0006c71b05ba900b55b24a670f35b3c7271e62f80ca"} Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.775325 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerStarted","Data":"0193e4873f3244cab191ab4be299640f973319dfcf035f612a78cdfcd54c17a7"} Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.775370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerStarted","Data":"8ce8e611661071a5e49fc120f1a9bef7f615630c825216f4eed7d8cb555e5d6c"} Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.807753 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.8077245939999997 podStartE2EDuration="3.807724594s" podCreationTimestamp="2025-12-09 11:56:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:16.801767892 +0000 UTC m=+1502.042526431" watchObservedRunningTime="2025-12-09 11:56:16.807724594 +0000 UTC m=+1502.048483123" Dec 09 11:56:16 crc kubenswrapper[4770]: I1209 11:56:16.825519 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.825494435 podStartE2EDuration="3.825494435s" podCreationTimestamp="2025-12-09 11:56:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:16.822663033 +0000 UTC m=+1502.063421552" watchObservedRunningTime="2025-12-09 11:56:16.825494435 +0000 UTC m=+1502.066252954" Dec 09 11:56:18 crc kubenswrapper[4770]: I1209 11:56:18.748332 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 11:56:18 crc kubenswrapper[4770]: I1209 11:56:18.752643 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 11:56:23 crc kubenswrapper[4770]: I1209 11:56:23.852112 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerID="0ca6ad8ec1507c0162d39456fba3bac7cf36c9d4e3c5e8ad7d0894c5c5cc16b6" exitCode=137 Dec 09 11:56:23 crc kubenswrapper[4770]: I1209 11:56:23.852328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerDied","Data":"0ca6ad8ec1507c0162d39456fba3bac7cf36c9d4e3c5e8ad7d0894c5c5cc16b6"} Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.197692 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.197742 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.238258 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.252531 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.256207 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.256274 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.302251 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.315631 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.732780 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.853107 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjwc9\" (UniqueName: \"kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.853203 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.853311 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.853461 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.854414 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.854496 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.854946 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.855035 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.855058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle\") pod \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\" (UID: \"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5\") " Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.856330 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.856697 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.860498 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts" (OuterVolumeSpecName: "scripts") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.869077 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4f6e95d-7134-4d48-99d6-17f9a8aab9a5","Type":"ContainerDied","Data":"ced6a426d1230370028e4de570483002be32e6a2bca00d23176595cd237e62a0"} Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.869153 4770 scope.go:117] "RemoveContainer" containerID="0ca6ad8ec1507c0162d39456fba3bac7cf36c9d4e3c5e8ad7d0894c5c5cc16b6" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.869237 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.871456 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.871485 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.871500 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.871544 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.887039 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9" (OuterVolumeSpecName: "kube-api-access-vjwc9") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "kube-api-access-vjwc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.901223 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.959347 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjwc9\" (UniqueName: \"kubernetes.io/projected/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-kube-api-access-vjwc9\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.959398 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:24 crc kubenswrapper[4770]: I1209 11:56:24.959414 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.011651 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.052107 4770 scope.go:117] "RemoveContainer" containerID="f55f463367ab1c8bc35972b6a41d9cb825b48d3909ce0451765716ca6c0a0267" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.053960 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data" (OuterVolumeSpecName: "config-data") pod "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" (UID: "e4f6e95d-7134-4d48-99d6-17f9a8aab9a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.061391 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.061432 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.088187 4770 scope.go:117] "RemoveContainer" containerID="89f03724fff6c534e61057eb4e541ada12463fbab10e60ff30ca842ba4a9c22c" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.147298 4770 scope.go:117] "RemoveContainer" containerID="af8f2cb0c684d4a22cfce2984e7a27e5e53dab1913cc27351bef3652477fae17" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.221701 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.234066 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.245971 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:56:25 crc kubenswrapper[4770]: E1209 11:56:25.246532 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-notification-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246553 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-notification-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: E1209 11:56:25.246572 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="sg-core" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246579 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="sg-core" Dec 09 11:56:25 crc kubenswrapper[4770]: E1209 11:56:25.246616 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="proxy-httpd" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246626 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="proxy-httpd" Dec 09 11:56:25 crc kubenswrapper[4770]: E1209 11:56:25.246658 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-central-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246666 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-central-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246944 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-central-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246970 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="sg-core" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246987 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="proxy-httpd" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.246999 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" containerName="ceilometer-notification-agent" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.249576 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.265460 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.265934 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.306973 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.356294 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f6e95d-7134-4d48-99d6-17f9a8aab9a5" path="/var/lib/kubelet/pods/e4f6e95d-7134-4d48-99d6-17f9a8aab9a5/volumes" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369200 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clp74\" (UniqueName: \"kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369272 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369294 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369774 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369800 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.369881 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.370159 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: E1209 11:56:25.451940 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f6e95d_7134_4d48_99d6_17f9a8aab9a5.slice/crio-ced6a426d1230370028e4de570483002be32e6a2bca00d23176595cd237e62a0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f6e95d_7134_4d48_99d6_17f9a8aab9a5.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.471894 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472174 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472334 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472451 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472572 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472700 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.472819 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clp74\" (UniqueName: \"kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.474071 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.474191 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.478516 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.478547 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.478763 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.481456 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.491541 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clp74\" (UniqueName: \"kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74\") pod \"ceilometer-0\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " pod="openstack/ceilometer-0" Dec 09 11:56:25 crc kubenswrapper[4770]: I1209 11:56:25.595996 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:56:26 crc kubenswrapper[4770]: I1209 11:56:26.106843 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:56:26 crc kubenswrapper[4770]: I1209 11:56:26.899065 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerStarted","Data":"9ffcf624f2c2a4435893f7febc213e847b94ffe8fe9970fcb45fd5a3708f7cd2"} Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.419012 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.419472 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.425177 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.491991 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.492105 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.903919 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 11:56:27 crc kubenswrapper[4770]: I1209 11:56:27.910456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerStarted","Data":"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c"} Dec 09 11:56:30 crc kubenswrapper[4770]: I1209 11:56:30.942743 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerStarted","Data":"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30"} Dec 09 11:56:31 crc kubenswrapper[4770]: I1209 11:56:31.956216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerStarted","Data":"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0"} Dec 09 11:56:33 crc kubenswrapper[4770]: I1209 11:56:33.977151 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerStarted","Data":"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42"} Dec 09 11:56:33 crc kubenswrapper[4770]: I1209 11:56:33.977722 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 11:56:34 crc kubenswrapper[4770]: I1209 11:56:34.007769 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.229656837 podStartE2EDuration="9.007746996s" podCreationTimestamp="2025-12-09 11:56:25 +0000 UTC" firstStartedPulling="2025-12-09 11:56:26.122712189 +0000 UTC m=+1511.363470708" lastFinishedPulling="2025-12-09 11:56:32.900802348 +0000 UTC m=+1518.141560867" observedRunningTime="2025-12-09 11:56:34.001418435 +0000 UTC m=+1519.242176974" watchObservedRunningTime="2025-12-09 11:56:34.007746996 +0000 UTC m=+1519.248505505" Dec 09 11:56:37 crc kubenswrapper[4770]: I1209 11:56:37.009596 4770 generic.go:334] "Generic (PLEG): container finished" podID="eb132594-6dbe-452f-ab35-8337509cd95d" containerID="f483d04a4a2d6845118ac8581f5705bae9286fde5e96eb516d573219ca6e36da" exitCode=0 Dec 09 11:56:37 crc kubenswrapper[4770]: I1209 11:56:37.009938 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8twmk" event={"ID":"eb132594-6dbe-452f-ab35-8337509cd95d","Type":"ContainerDied","Data":"f483d04a4a2d6845118ac8581f5705bae9286fde5e96eb516d573219ca6e36da"} Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.440875 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.556871 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data\") pod \"eb132594-6dbe-452f-ab35-8337509cd95d\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.557075 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts\") pod \"eb132594-6dbe-452f-ab35-8337509cd95d\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.557113 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle\") pod \"eb132594-6dbe-452f-ab35-8337509cd95d\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.557322 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blpgv\" (UniqueName: \"kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv\") pod \"eb132594-6dbe-452f-ab35-8337509cd95d\" (UID: \"eb132594-6dbe-452f-ab35-8337509cd95d\") " Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.563464 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts" (OuterVolumeSpecName: "scripts") pod "eb132594-6dbe-452f-ab35-8337509cd95d" (UID: "eb132594-6dbe-452f-ab35-8337509cd95d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.563979 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv" (OuterVolumeSpecName: "kube-api-access-blpgv") pod "eb132594-6dbe-452f-ab35-8337509cd95d" (UID: "eb132594-6dbe-452f-ab35-8337509cd95d"). InnerVolumeSpecName "kube-api-access-blpgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.600263 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data" (OuterVolumeSpecName: "config-data") pod "eb132594-6dbe-452f-ab35-8337509cd95d" (UID: "eb132594-6dbe-452f-ab35-8337509cd95d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.603611 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb132594-6dbe-452f-ab35-8337509cd95d" (UID: "eb132594-6dbe-452f-ab35-8337509cd95d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.660181 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.660221 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.660241 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blpgv\" (UniqueName: \"kubernetes.io/projected/eb132594-6dbe-452f-ab35-8337509cd95d-kube-api-access-blpgv\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:38 crc kubenswrapper[4770]: I1209 11:56:38.660253 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb132594-6dbe-452f-ab35-8337509cd95d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.030642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8twmk" event={"ID":"eb132594-6dbe-452f-ab35-8337509cd95d","Type":"ContainerDied","Data":"796c18a5311419f84f129ca3ce4aba68e8a6be7c1c17d1f6ddf11958587ebf38"} Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.030686 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8twmk" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.030695 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="796c18a5311419f84f129ca3ce4aba68e8a6be7c1c17d1f6ddf11958587ebf38" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.134743 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:56:39 crc kubenswrapper[4770]: E1209 11:56:39.135441 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb132594-6dbe-452f-ab35-8337509cd95d" containerName="nova-cell0-conductor-db-sync" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.135567 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb132594-6dbe-452f-ab35-8337509cd95d" containerName="nova-cell0-conductor-db-sync" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.135884 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb132594-6dbe-452f-ab35-8337509cd95d" containerName="nova-cell0-conductor-db-sync" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.136684 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.139503 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.139736 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gdt78" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.148773 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.170556 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmcg6\" (UniqueName: \"kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.170614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.170666 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.272855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmcg6\" (UniqueName: \"kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.273342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.273512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.277713 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.288166 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.293887 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmcg6\" (UniqueName: \"kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6\") pod \"nova-cell0-conductor-0\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.457172 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:39 crc kubenswrapper[4770]: I1209 11:56:39.912912 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:56:40 crc kubenswrapper[4770]: I1209 11:56:40.042670 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e77fad7-c892-4d17-87ca-6c699f1b2258","Type":"ContainerStarted","Data":"d466c3774d7cf034d786a2d6a4364f0240ad291c945c2f9e6d0322daf6b02e04"} Dec 09 11:56:41 crc kubenswrapper[4770]: I1209 11:56:41.053093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e77fad7-c892-4d17-87ca-6c699f1b2258","Type":"ContainerStarted","Data":"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b"} Dec 09 11:56:41 crc kubenswrapper[4770]: I1209 11:56:41.053756 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:49 crc kubenswrapper[4770]: I1209 11:56:49.484172 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 09 11:56:49 crc kubenswrapper[4770]: I1209 11:56:49.519658 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.519630353 podStartE2EDuration="10.519630353s" podCreationTimestamp="2025-12-09 11:56:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:41.080830925 +0000 UTC m=+1526.321589444" watchObservedRunningTime="2025-12-09 11:56:49.519630353 +0000 UTC m=+1534.760388872" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.080560 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jb49v"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.082683 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.084815 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.090406 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.094999 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jb49v"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.248531 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7x5n\" (UniqueName: \"kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.248601 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.248665 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.248766 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.350502 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.351054 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.351309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7x5n\" (UniqueName: \"kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.351444 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.359415 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.359634 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.359789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.381595 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7x5n\" (UniqueName: \"kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n\") pod \"nova-cell0-cell-mapping-jb49v\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.386142 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.388118 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.391193 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.400243 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.409865 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.497073 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.500035 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.502346 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.518199 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.556461 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.556637 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.556664 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tf5s\" (UniqueName: \"kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.556690 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.620986 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.623092 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.625652 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.659747 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.659870 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v5bt\" (UniqueName: \"kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.659929 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.663847 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.668035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.668079 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tf5s\" (UniqueName: \"kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.668122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.668249 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.670824 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.671957 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.675997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.687579 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.694336 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.706507 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.730551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tf5s\" (UniqueName: \"kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s\") pod \"nova-api-0\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.733993 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.771755 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v5bt\" (UniqueName: \"kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.771832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.771893 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.771976 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.772065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52ggm\" (UniqueName: \"kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.772167 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.781703 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.782986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.807378 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v5bt\" (UniqueName: \"kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.829172 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.831359 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.866552 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.895294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.895735 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.895867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.896018 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.896250 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52ggm\" (UniqueName: \"kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.896570 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw6h9\" (UniqueName: \"kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.896712 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.897493 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.909590 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.922611 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.943432 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52ggm\" (UniqueName: \"kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:50 crc kubenswrapper[4770]: I1209 11:56:50.953412 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data\") pod \"nova-scheduler-0\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " pod="openstack/nova-scheduler-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:50.999391 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:50.999947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000010 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000064 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwl65\" (UniqueName: \"kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000107 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000137 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000189 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000293 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw6h9\" (UniqueName: \"kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000332 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000358 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.000486 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.008267 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.009781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.034093 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw6h9\" (UniqueName: \"kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9\") pod \"nova-metadata-0\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.065506 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jb49v"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.085198 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109255 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109719 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109784 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwl65\" (UniqueName: \"kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109829 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.109861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.111322 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.113718 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.114506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.114570 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.115150 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.117175 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.133593 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwl65\" (UniqueName: \"kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65\") pod \"dnsmasq-dns-5bfb54f9b5-tk688\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.160697 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.214679 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jb49v" event={"ID":"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a","Type":"ContainerStarted","Data":"228c30b14229e07d101d79a85caa91ab8b0e1ddc9123f1e1821dbc7b7fe19dad"} Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.385398 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qdtv7"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.387405 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qdtv7"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.387512 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.390458 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.390797 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.522441 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.522521 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.522713 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsh5j\" (UniqueName: \"kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.522762 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.528772 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.622265 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.625404 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.625450 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.625516 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsh5j\" (UniqueName: \"kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.625560 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.638272 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.649285 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.650864 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.666755 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsh5j\" (UniqueName: \"kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j\") pod \"nova-cell1-conductor-db-sync-qdtv7\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.806675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.816767 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:56:51 crc kubenswrapper[4770]: I1209 11:56:51.829255 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.029214 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.243370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jb49v" event={"ID":"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a","Type":"ContainerStarted","Data":"b221dc48fbe0244448017160d5eb9b96fc984cc9361787e38ba36c7d6f5633ab"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.251626 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e22b640-176b-4c5e-9747-190888df493a","Type":"ContainerStarted","Data":"40c241d9eedc435ff52f3efce46fa328324ebb3da3844a9c4e4408abf8000ee9"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.253655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" event={"ID":"07fb6c2a-e734-4169-9f47-c093a215f9ec","Type":"ContainerStarted","Data":"9416370cef5624f158deb7862fc13e2e987801a42abf44957346b6d155484d22"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.256915 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc5b7374-97a8-44a5-9537-da68c81a0533","Type":"ContainerStarted","Data":"c811bfc998c602ad433caaa5374d218b57ae82e4b16278b39bd1c72dfceea560"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.263369 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerStarted","Data":"8412495736e7a143e045411a2faab093f66ed3898e72b72af5d6ed7cc0e0a266"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.265615 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerStarted","Data":"5d49cd0f59e6f05376f2d7efab34348333f57b3939acd515cdae46de4aa46cbb"} Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.266506 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jb49v" podStartSLOduration=2.2664835500000002 podStartE2EDuration="2.26648355s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:52.26173927 +0000 UTC m=+1537.502497799" watchObservedRunningTime="2025-12-09 11:56:52.26648355 +0000 UTC m=+1537.507242069" Dec 09 11:56:52 crc kubenswrapper[4770]: I1209 11:56:52.390013 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qdtv7"] Dec 09 11:56:52 crc kubenswrapper[4770]: W1209 11:56:52.394353 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5dfb341_2abb_4794_b559_1bee29b28016.slice/crio-a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7 WatchSource:0}: Error finding container a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7: Status 404 returned error can't find the container with id a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7 Dec 09 11:56:53 crc kubenswrapper[4770]: I1209 11:56:53.284265 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" event={"ID":"b5dfb341-2abb-4794-b559-1bee29b28016","Type":"ContainerStarted","Data":"1292e06e7f5b284f4f90cb329ddfb73c6b447dce5d6ada412db691e074d9d6e8"} Dec 09 11:56:53 crc kubenswrapper[4770]: I1209 11:56:53.284560 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" event={"ID":"b5dfb341-2abb-4794-b559-1bee29b28016","Type":"ContainerStarted","Data":"a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7"} Dec 09 11:56:53 crc kubenswrapper[4770]: I1209 11:56:53.289780 4770 generic.go:334] "Generic (PLEG): container finished" podID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerID="73acd5b0d78a865d73b43f88829aa4a826410576bc0db186ae87a7c400450bbb" exitCode=0 Dec 09 11:56:53 crc kubenswrapper[4770]: I1209 11:56:53.291116 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" event={"ID":"07fb6c2a-e734-4169-9f47-c093a215f9ec","Type":"ContainerDied","Data":"73acd5b0d78a865d73b43f88829aa4a826410576bc0db186ae87a7c400450bbb"} Dec 09 11:56:53 crc kubenswrapper[4770]: I1209 11:56:53.313324 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" podStartSLOduration=2.31328206 podStartE2EDuration="2.31328206s" podCreationTimestamp="2025-12-09 11:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:53.303991184 +0000 UTC m=+1538.544749703" watchObservedRunningTime="2025-12-09 11:56:53.31328206 +0000 UTC m=+1538.554040589" Dec 09 11:56:55 crc kubenswrapper[4770]: I1209 11:56:55.000217 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:56:55 crc kubenswrapper[4770]: I1209 11:56:55.021265 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:56:55 crc kubenswrapper[4770]: I1209 11:56:55.613229 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.378802 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e22b640-176b-4c5e-9747-190888df493a","Type":"ContainerStarted","Data":"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.378984 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5e22b640-176b-4c5e-9747-190888df493a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341" gracePeriod=30 Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.382433 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" event={"ID":"07fb6c2a-e734-4169-9f47-c093a215f9ec","Type":"ContainerStarted","Data":"e143040a64c122594f975a4e26cc8fc4b0b71048af4090cbf1c907cc2b1720dc"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.382575 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.384813 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc5b7374-97a8-44a5-9537-da68c81a0533","Type":"ContainerStarted","Data":"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.387482 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-log" containerID="cri-o://6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a" gracePeriod=30 Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.387786 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerStarted","Data":"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.387811 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerStarted","Data":"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.387857 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-metadata" containerID="cri-o://1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba" gracePeriod=30 Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.399393 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.36763291 podStartE2EDuration="6.399375053s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="2025-12-09 11:56:51.679181718 +0000 UTC m=+1536.919940237" lastFinishedPulling="2025-12-09 11:56:55.710923861 +0000 UTC m=+1540.951682380" observedRunningTime="2025-12-09 11:56:56.395097524 +0000 UTC m=+1541.635856043" watchObservedRunningTime="2025-12-09 11:56:56.399375053 +0000 UTC m=+1541.640133572" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.407923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerStarted","Data":"82557840fb376c018845c2ef556557d7bd2bdd06d4eebe4296c7b929eee28b55"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.407977 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerStarted","Data":"88cbccbcb372a92b554fc1b323246de00ed0703fe0f0670fef15cd010dc24750"} Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.430158 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.619888664 podStartE2EDuration="6.430136714s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="2025-12-09 11:56:51.864977155 +0000 UTC m=+1537.105735674" lastFinishedPulling="2025-12-09 11:56:55.675225205 +0000 UTC m=+1540.915983724" observedRunningTime="2025-12-09 11:56:56.412847535 +0000 UTC m=+1541.653606054" watchObservedRunningTime="2025-12-09 11:56:56.430136714 +0000 UTC m=+1541.670895233" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.495690 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" podStartSLOduration=6.495464983 podStartE2EDuration="6.495464983s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:56:56.440372594 +0000 UTC m=+1541.681131113" watchObservedRunningTime="2025-12-09 11:56:56.495464983 +0000 UTC m=+1541.736223502" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.509864 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.699620029 podStartE2EDuration="6.509840578s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="2025-12-09 11:56:51.864979065 +0000 UTC m=+1537.105737584" lastFinishedPulling="2025-12-09 11:56:55.675199614 +0000 UTC m=+1540.915958133" observedRunningTime="2025-12-09 11:56:56.486260589 +0000 UTC m=+1541.727019108" watchObservedRunningTime="2025-12-09 11:56:56.509840578 +0000 UTC m=+1541.750599097" Dec 09 11:56:56 crc kubenswrapper[4770]: I1209 11:56:56.533762 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.367749382 podStartE2EDuration="6.533731574s" podCreationTimestamp="2025-12-09 11:56:50 +0000 UTC" firstStartedPulling="2025-12-09 11:56:51.544157499 +0000 UTC m=+1536.784916018" lastFinishedPulling="2025-12-09 11:56:55.710139691 +0000 UTC m=+1540.950898210" observedRunningTime="2025-12-09 11:56:56.529061986 +0000 UTC m=+1541.769820515" watchObservedRunningTime="2025-12-09 11:56:56.533731574 +0000 UTC m=+1541.774490093" Dec 09 11:56:57 crc kubenswrapper[4770]: I1209 11:56:57.432264 4770 generic.go:334] "Generic (PLEG): container finished" podID="84ade006-1109-4d4f-94f5-600457a7e178" containerID="6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a" exitCode=143 Dec 09 11:56:57 crc kubenswrapper[4770]: I1209 11:56:57.432507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerDied","Data":"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a"} Dec 09 11:57:00 crc kubenswrapper[4770]: I1209 11:57:00.452327 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:00 crc kubenswrapper[4770]: I1209 11:57:00.453374 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" containerName="kube-state-metrics" containerID="cri-o://318e973a919a8e953fba7ee24962721c2b32ac548f3e70514be2e873a194e57c" gracePeriod=30 Dec 09 11:57:00 crc kubenswrapper[4770]: I1209 11:57:00.897861 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:57:00 crc kubenswrapper[4770]: I1209 11:57:00.901798 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:57:00 crc kubenswrapper[4770]: I1209 11:57:00.923968 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.089300 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.089372 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.118034 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.118110 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.172286 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.284714 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.285161 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="dnsmasq-dns" containerID="cri-o://d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e" gracePeriod=10 Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.315587 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.479574 4770 generic.go:334] "Generic (PLEG): container finished" podID="76483913-5fec-49d0-8d19-916a024ec7fd" containerID="318e973a919a8e953fba7ee24962721c2b32ac548f3e70514be2e873a194e57c" exitCode=2 Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.479642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76483913-5fec-49d0-8d19-916a024ec7fd","Type":"ContainerDied","Data":"318e973a919a8e953fba7ee24962721c2b32ac548f3e70514be2e873a194e57c"} Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.490782 4770 generic.go:334] "Generic (PLEG): container finished" podID="0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" containerID="b221dc48fbe0244448017160d5eb9b96fc984cc9361787e38ba36c7d6f5633ab" exitCode=0 Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.491928 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jb49v" event={"ID":"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a","Type":"ContainerDied","Data":"b221dc48fbe0244448017160d5eb9b96fc984cc9361787e38ba36c7d6f5633ab"} Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.607349 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.944813 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.980197 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:01 crc kubenswrapper[4770]: I1209 11:57:01.980318 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.114020 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.114827 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qqwc\" (UniqueName: \"kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc\") pod \"76483913-5fec-49d0-8d19-916a024ec7fd\" (UID: \"76483913-5fec-49d0-8d19-916a024ec7fd\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.144362 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc" (OuterVolumeSpecName: "kube-api-access-6qqwc") pod "76483913-5fec-49d0-8d19-916a024ec7fd" (UID: "76483913-5fec-49d0-8d19-916a024ec7fd"). InnerVolumeSpecName "kube-api-access-6qqwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217483 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217597 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrkc4\" (UniqueName: \"kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217693 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217731 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.217819 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc\") pod \"a21d330f-a9b8-49b7-a035-83c9215ced97\" (UID: \"a21d330f-a9b8-49b7-a035-83c9215ced97\") " Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.218695 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qqwc\" (UniqueName: \"kubernetes.io/projected/76483913-5fec-49d0-8d19-916a024ec7fd-kube-api-access-6qqwc\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.235329 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4" (OuterVolumeSpecName: "kube-api-access-mrkc4") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "kube-api-access-mrkc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.290627 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.312553 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config" (OuterVolumeSpecName: "config") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.312568 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.321416 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.321470 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrkc4\" (UniqueName: \"kubernetes.io/projected/a21d330f-a9b8-49b7-a035-83c9215ced97-kube-api-access-mrkc4\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.321509 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.321522 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.337640 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.369558 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a21d330f-a9b8-49b7-a035-83c9215ced97" (UID: "a21d330f-a9b8-49b7-a035-83c9215ced97"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.423955 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.424008 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a21d330f-a9b8-49b7-a035-83c9215ced97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.503292 4770 generic.go:334] "Generic (PLEG): container finished" podID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerID="d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e" exitCode=0 Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.503359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" event={"ID":"a21d330f-a9b8-49b7-a035-83c9215ced97","Type":"ContainerDied","Data":"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e"} Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.503425 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" event={"ID":"a21d330f-a9b8-49b7-a035-83c9215ced97","Type":"ContainerDied","Data":"d4d34001c16ef9075ef6bb2cf1967166be3051826e2fc29771a0d9154eafb82f"} Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.503379 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4f5fc4f-2lgx7" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.503457 4770 scope.go:117] "RemoveContainer" containerID="d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.510989 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.516758 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76483913-5fec-49d0-8d19-916a024ec7fd","Type":"ContainerDied","Data":"0010a467df1fe6e371c53ff6a681be6a0e157914c226d73cb140cda8dd5d73cc"} Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.530119 4770 scope.go:117] "RemoveContainer" containerID="13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.573918 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.606246 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b4f5fc4f-2lgx7"] Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.637988 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.639074 4770 scope.go:117] "RemoveContainer" containerID="d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e" Dec 09 11:57:02 crc kubenswrapper[4770]: E1209 11:57:02.648296 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e\": container with ID starting with d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e not found: ID does not exist" containerID="d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.648380 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e"} err="failed to get container status \"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e\": rpc error: code = NotFound desc = could not find container \"d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e\": container with ID starting with d3324bd670c827d5338ccef7d252298081f4cc586a0abb447c9398998da47c0e not found: ID does not exist" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.648418 4770 scope.go:117] "RemoveContainer" containerID="13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9" Dec 09 11:57:02 crc kubenswrapper[4770]: E1209 11:57:02.648956 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9\": container with ID starting with 13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9 not found: ID does not exist" containerID="13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.649008 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9"} err="failed to get container status \"13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9\": rpc error: code = NotFound desc = could not find container \"13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9\": container with ID starting with 13df5e1dd8083507b56a78abc1622341f3363d8761cbcba6dc75dd5e2f8524e9 not found: ID does not exist" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.649039 4770 scope.go:117] "RemoveContainer" containerID="318e973a919a8e953fba7ee24962721c2b32ac548f3e70514be2e873a194e57c" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.651236 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.677986 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:02 crc kubenswrapper[4770]: E1209 11:57:02.678559 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="dnsmasq-dns" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.678578 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="dnsmasq-dns" Dec 09 11:57:02 crc kubenswrapper[4770]: E1209 11:57:02.678610 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="init" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.678616 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="init" Dec 09 11:57:02 crc kubenswrapper[4770]: E1209 11:57:02.678655 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" containerName="kube-state-metrics" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.678666 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" containerName="kube-state-metrics" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.679067 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" containerName="dnsmasq-dns" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.679103 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" containerName="kube-state-metrics" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.680054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.683309 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.684358 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.689081 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.832221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258z8\" (UniqueName: \"kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.832557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.832618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.832830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.934460 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258z8\" (UniqueName: \"kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.934537 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.934618 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.934705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.939306 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.939306 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.956183 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258z8\" (UniqueName: \"kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:02 crc kubenswrapper[4770]: I1209 11:57:02.958466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " pod="openstack/kube-state-metrics-0" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.008585 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.046553 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.139197 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts\") pod \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.139385 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data\") pod \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.139564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle\") pod \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.139650 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7x5n\" (UniqueName: \"kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n\") pod \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\" (UID: \"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a\") " Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.144972 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts" (OuterVolumeSpecName: "scripts") pod "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" (UID: "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.149813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n" (OuterVolumeSpecName: "kube-api-access-r7x5n") pod "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" (UID: "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a"). InnerVolumeSpecName "kube-api-access-r7x5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.193113 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" (UID: "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.237064 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data" (OuterVolumeSpecName: "config-data") pod "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" (UID: "0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.254340 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.254396 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7x5n\" (UniqueName: \"kubernetes.io/projected/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-kube-api-access-r7x5n\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.254408 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.254418 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.363640 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76483913-5fec-49d0-8d19-916a024ec7fd" path="/var/lib/kubelet/pods/76483913-5fec-49d0-8d19-916a024ec7fd/volumes" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.374964 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a21d330f-a9b8-49b7-a035-83c9215ced97" path="/var/lib/kubelet/pods/a21d330f-a9b8-49b7-a035-83c9215ced97/volumes" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.601465 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jb49v" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.603424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jb49v" event={"ID":"0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a","Type":"ContainerDied","Data":"228c30b14229e07d101d79a85caa91ab8b0e1ddc9123f1e1821dbc7b7fe19dad"} Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.603459 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="228c30b14229e07d101d79a85caa91ab8b0e1ddc9123f1e1821dbc7b7fe19dad" Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.659242 4770 generic.go:334] "Generic (PLEG): container finished" podID="b5dfb341-2abb-4794-b559-1bee29b28016" containerID="1292e06e7f5b284f4f90cb329ddfb73c6b447dce5d6ada412db691e074d9d6e8" exitCode=0 Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.659304 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" event={"ID":"b5dfb341-2abb-4794-b559-1bee29b28016","Type":"ContainerDied","Data":"1292e06e7f5b284f4f90cb329ddfb73c6b447dce5d6ada412db691e074d9d6e8"} Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.686931 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.871728 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.872106 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-log" containerID="cri-o://88cbccbcb372a92b554fc1b323246de00ed0703fe0f0670fef15cd010dc24750" gracePeriod=30 Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.872623 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-api" containerID="cri-o://82557840fb376c018845c2ef556557d7bd2bdd06d4eebe4296c7b929eee28b55" gracePeriod=30 Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.892791 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:03 crc kubenswrapper[4770]: I1209 11:57:03.893099 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerName="nova-scheduler-scheduler" containerID="cri-o://e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" gracePeriod=30 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.647380 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.648070 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-central-agent" containerID="cri-o://459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c" gracePeriod=30 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.648139 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="sg-core" containerID="cri-o://7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0" gracePeriod=30 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.648154 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-notification-agent" containerID="cri-o://b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30" gracePeriod=30 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.648142 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="proxy-httpd" containerID="cri-o://f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42" gracePeriod=30 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.671634 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b","Type":"ContainerStarted","Data":"218393e2da0bbfb5410ec4757c8cb027246a3f4ac476769670c88ceb23a30731"} Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.671702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b","Type":"ContainerStarted","Data":"a9346dc03c35e03914390456553753b562c2a058abe52edb39c068776fc9b55b"} Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.671721 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.674172 4770 generic.go:334] "Generic (PLEG): container finished" podID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerID="88cbccbcb372a92b554fc1b323246de00ed0703fe0f0670fef15cd010dc24750" exitCode=143 Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.674322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerDied","Data":"88cbccbcb372a92b554fc1b323246de00ed0703fe0f0670fef15cd010dc24750"} Dec 09 11:57:04 crc kubenswrapper[4770]: I1209 11:57:04.689425 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.160689027 podStartE2EDuration="2.689404612s" podCreationTimestamp="2025-12-09 11:57:02 +0000 UTC" firstStartedPulling="2025-12-09 11:57:03.635864361 +0000 UTC m=+1548.876622880" lastFinishedPulling="2025-12-09 11:57:04.164579936 +0000 UTC m=+1549.405338465" observedRunningTime="2025-12-09 11:57:04.686645232 +0000 UTC m=+1549.927403751" watchObservedRunningTime="2025-12-09 11:57:04.689404612 +0000 UTC m=+1549.930163131" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.191909 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.286067 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts\") pod \"b5dfb341-2abb-4794-b559-1bee29b28016\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.286129 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data\") pod \"b5dfb341-2abb-4794-b559-1bee29b28016\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.286209 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsh5j\" (UniqueName: \"kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j\") pod \"b5dfb341-2abb-4794-b559-1bee29b28016\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.286332 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle\") pod \"b5dfb341-2abb-4794-b559-1bee29b28016\" (UID: \"b5dfb341-2abb-4794-b559-1bee29b28016\") " Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.292278 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts" (OuterVolumeSpecName: "scripts") pod "b5dfb341-2abb-4794-b559-1bee29b28016" (UID: "b5dfb341-2abb-4794-b559-1bee29b28016"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.297217 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j" (OuterVolumeSpecName: "kube-api-access-gsh5j") pod "b5dfb341-2abb-4794-b559-1bee29b28016" (UID: "b5dfb341-2abb-4794-b559-1bee29b28016"). InnerVolumeSpecName "kube-api-access-gsh5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.318155 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data" (OuterVolumeSpecName: "config-data") pod "b5dfb341-2abb-4794-b559-1bee29b28016" (UID: "b5dfb341-2abb-4794-b559-1bee29b28016"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.320023 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5dfb341-2abb-4794-b559-1bee29b28016" (UID: "b5dfb341-2abb-4794-b559-1bee29b28016"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.388492 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.388539 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.388553 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5dfb341-2abb-4794-b559-1bee29b28016-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.388564 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsh5j\" (UniqueName: \"kubernetes.io/projected/b5dfb341-2abb-4794-b559-1bee29b28016-kube-api-access-gsh5j\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.685651 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.685642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qdtv7" event={"ID":"b5dfb341-2abb-4794-b559-1bee29b28016","Type":"ContainerDied","Data":"a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7"} Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.685707 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1216fa05977d82bc226ef939ade7024f6b60965304bdbe9b17ebb33098715b7" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689749 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerID="f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42" exitCode=0 Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689780 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerID="7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0" exitCode=2 Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689791 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerID="459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c" exitCode=0 Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689839 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerDied","Data":"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42"} Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerDied","Data":"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0"} Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.689889 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerDied","Data":"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c"} Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.788141 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:57:05 crc kubenswrapper[4770]: E1209 11:57:05.788842 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" containerName="nova-manage" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.788867 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" containerName="nova-manage" Dec 09 11:57:05 crc kubenswrapper[4770]: E1209 11:57:05.788937 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5dfb341-2abb-4794-b559-1bee29b28016" containerName="nova-cell1-conductor-db-sync" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.788946 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5dfb341-2abb-4794-b559-1bee29b28016" containerName="nova-cell1-conductor-db-sync" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.789193 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" containerName="nova-manage" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.789226 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5dfb341-2abb-4794-b559-1bee29b28016" containerName="nova-cell1-conductor-db-sync" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.790068 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.795832 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxkcb\" (UniqueName: \"kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.795894 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.796316 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.796409 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.798035 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.898842 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.898920 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.899005 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxkcb\" (UniqueName: \"kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.924783 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.933863 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:05 crc kubenswrapper[4770]: I1209 11:57:05.940474 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxkcb\" (UniqueName: \"kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb\") pod \"nova-cell1-conductor-0\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:06 crc kubenswrapper[4770]: I1209 11:57:06.011518 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:06 crc kubenswrapper[4770]: E1209 11:57:06.105359 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:06 crc kubenswrapper[4770]: E1209 11:57:06.109245 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:06 crc kubenswrapper[4770]: E1209 11:57:06.110767 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:06 crc kubenswrapper[4770]: E1209 11:57:06.110835 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerName="nova-scheduler-scheduler" Dec 09 11:57:06 crc kubenswrapper[4770]: I1209 11:57:06.547465 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:57:06 crc kubenswrapper[4770]: I1209 11:57:06.709269 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"df40ca35-a10e-4571-b460-f7cd465070d8","Type":"ContainerStarted","Data":"b0f18994898aaa4cc5d7d372aa4ba2389d90e5d2cae9bac4e5de4333b9e7c985"} Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.603233 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647253 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647435 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647492 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clp74\" (UniqueName: \"kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647598 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647691 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647721 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.647754 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml\") pod \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\" (UID: \"6f75b1d7-1813-4abf-a408-f0e58518e0a2\") " Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.648145 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.648185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.648525 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.648547 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f75b1d7-1813-4abf-a408-f0e58518e0a2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.657137 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts" (OuterVolumeSpecName: "scripts") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.657181 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74" (OuterVolumeSpecName: "kube-api-access-clp74") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "kube-api-access-clp74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.702731 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.731410 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"df40ca35-a10e-4571-b460-f7cd465070d8","Type":"ContainerStarted","Data":"5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6"} Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.736068 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.747827 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerID="b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30" exitCode=0 Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.747917 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerDied","Data":"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30"} Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.747943 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f75b1d7-1813-4abf-a408-f0e58518e0a2","Type":"ContainerDied","Data":"9ffcf624f2c2a4435893f7febc213e847b94ffe8fe9970fcb45fd5a3708f7cd2"} Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.747962 4770 scope.go:117] "RemoveContainer" containerID="f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.748106 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.752143 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.752195 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clp74\" (UniqueName: \"kubernetes.io/projected/6f75b1d7-1813-4abf-a408-f0e58518e0a2-kube-api-access-clp74\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.752211 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.765063 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.765039058 podStartE2EDuration="2.765039058s" podCreationTimestamp="2025-12-09 11:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:07.757443165 +0000 UTC m=+1552.998201694" watchObservedRunningTime="2025-12-09 11:57:07.765039058 +0000 UTC m=+1553.005797577" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.795826 4770 scope.go:117] "RemoveContainer" containerID="7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.805715 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.817146 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data" (OuterVolumeSpecName: "config-data") pod "6f75b1d7-1813-4abf-a408-f0e58518e0a2" (UID: "6f75b1d7-1813-4abf-a408-f0e58518e0a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.854426 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.854486 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f75b1d7-1813-4abf-a408-f0e58518e0a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.899123 4770 scope.go:117] "RemoveContainer" containerID="b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.923684 4770 scope.go:117] "RemoveContainer" containerID="459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.943037 4770 scope.go:117] "RemoveContainer" containerID="f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42" Dec 09 11:57:07 crc kubenswrapper[4770]: E1209 11:57:07.943566 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42\": container with ID starting with f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42 not found: ID does not exist" containerID="f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.943608 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42"} err="failed to get container status \"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42\": rpc error: code = NotFound desc = could not find container \"f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42\": container with ID starting with f91ebd46d3904d22311d7d9166546754285cd1ea48f0d97a92bf71f19aaa3b42 not found: ID does not exist" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.943640 4770 scope.go:117] "RemoveContainer" containerID="7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0" Dec 09 11:57:07 crc kubenswrapper[4770]: E1209 11:57:07.944120 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0\": container with ID starting with 7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0 not found: ID does not exist" containerID="7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.944174 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0"} err="failed to get container status \"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0\": rpc error: code = NotFound desc = could not find container \"7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0\": container with ID starting with 7edd65a497825bae5547e0ea61eef7621e4b2758cdefbb386bf8a0f6dea299c0 not found: ID does not exist" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.944209 4770 scope.go:117] "RemoveContainer" containerID="b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30" Dec 09 11:57:07 crc kubenswrapper[4770]: E1209 11:57:07.944741 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30\": container with ID starting with b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30 not found: ID does not exist" containerID="b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.944785 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30"} err="failed to get container status \"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30\": rpc error: code = NotFound desc = could not find container \"b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30\": container with ID starting with b3b65ac330fb31992f7e75dd9bed74029c0026d644ce0450a90ff6f2f652bd30 not found: ID does not exist" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.944804 4770 scope.go:117] "RemoveContainer" containerID="459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c" Dec 09 11:57:07 crc kubenswrapper[4770]: E1209 11:57:07.945149 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c\": container with ID starting with 459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c not found: ID does not exist" containerID="459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c" Dec 09 11:57:07 crc kubenswrapper[4770]: I1209 11:57:07.945196 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c"} err="failed to get container status \"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c\": rpc error: code = NotFound desc = could not find container \"459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c\": container with ID starting with 459fdc9f67f01f2211674cc38e1b86eb6ddef1830d29d9e3a82d5522e0ada58c not found: ID does not exist" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.102726 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.116056 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.127693 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.128241 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="sg-core" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128264 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="sg-core" Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.128305 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-central-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128312 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-central-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.128324 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-notification-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128330 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-notification-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.128344 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="proxy-httpd" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128349 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="proxy-httpd" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128551 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-central-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128579 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="ceilometer-notification-agent" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128590 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="proxy-httpd" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.128596 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" containerName="sg-core" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.130617 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.133461 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.133598 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.136025 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.142351 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.158781 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.158843 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.158873 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.158887 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.158947 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.159193 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrj6c\" (UniqueName: \"kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.159302 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.159424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261196 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261267 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261331 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261395 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrj6c\" (UniqueName: \"kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261479 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261526 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261575 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.261645 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.262123 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.262747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.265676 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.266040 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.266626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.266844 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.276705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.282468 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrj6c\" (UniqueName: \"kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c\") pod \"ceilometer-0\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.456964 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.653306 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.668533 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle\") pod \"cc5b7374-97a8-44a5-9537-da68c81a0533\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.668813 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52ggm\" (UniqueName: \"kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm\") pod \"cc5b7374-97a8-44a5-9537-da68c81a0533\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.668887 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data\") pod \"cc5b7374-97a8-44a5-9537-da68c81a0533\" (UID: \"cc5b7374-97a8-44a5-9537-da68c81a0533\") " Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.681314 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm" (OuterVolumeSpecName: "kube-api-access-52ggm") pod "cc5b7374-97a8-44a5-9537-da68c81a0533" (UID: "cc5b7374-97a8-44a5-9537-da68c81a0533"). InnerVolumeSpecName "kube-api-access-52ggm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.714183 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data" (OuterVolumeSpecName: "config-data") pod "cc5b7374-97a8-44a5-9537-da68c81a0533" (UID: "cc5b7374-97a8-44a5-9537-da68c81a0533"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.732065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc5b7374-97a8-44a5-9537-da68c81a0533" (UID: "cc5b7374-97a8-44a5-9537-da68c81a0533"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.771158 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.771198 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc5b7374-97a8-44a5-9537-da68c81a0533-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.771214 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52ggm\" (UniqueName: \"kubernetes.io/projected/cc5b7374-97a8-44a5-9537-da68c81a0533-kube-api-access-52ggm\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.791151 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" exitCode=0 Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.791263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc5b7374-97a8-44a5-9537-da68c81a0533","Type":"ContainerDied","Data":"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb"} Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.791298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc5b7374-97a8-44a5-9537-da68c81a0533","Type":"ContainerDied","Data":"c811bfc998c602ad433caaa5374d218b57ae82e4b16278b39bd1c72dfceea560"} Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.791318 4770 scope.go:117] "RemoveContainer" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.791464 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.880190 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.900353 4770 scope.go:117] "RemoveContainer" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.900462 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.901291 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb\": container with ID starting with e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb not found: ID does not exist" containerID="e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.901362 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb"} err="failed to get container status \"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb\": rpc error: code = NotFound desc = could not find container \"e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb\": container with ID starting with e32c2645491d9b3b85587bfa036d3a7e4a3cd80190ed5f6184ca5c07bcafc9cb not found: ID does not exist" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.922516 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.947181 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: E1209 11:57:08.948021 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerName="nova-scheduler-scheduler" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.948077 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerName="nova-scheduler-scheduler" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.948427 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" containerName="nova-scheduler-scheduler" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.949712 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.952433 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.970139 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.976629 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.976717 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn8fw\" (UniqueName: \"kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:08 crc kubenswrapper[4770]: I1209 11:57:08.976763 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.077555 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn8fw\" (UniqueName: \"kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.077616 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.077758 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.087613 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.088303 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.099134 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn8fw\" (UniqueName: \"kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw\") pod \"nova-scheduler-0\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.276731 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.388106 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f75b1d7-1813-4abf-a408-f0e58518e0a2" path="/var/lib/kubelet/pods/6f75b1d7-1813-4abf-a408-f0e58518e0a2/volumes" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.389591 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc5b7374-97a8-44a5-9537-da68c81a0533" path="/var/lib/kubelet/pods/cc5b7374-97a8-44a5-9537-da68c81a0533/volumes" Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.847567 4770 generic.go:334] "Generic (PLEG): container finished" podID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerID="82557840fb376c018845c2ef556557d7bd2bdd06d4eebe4296c7b929eee28b55" exitCode=0 Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.848006 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerDied","Data":"82557840fb376c018845c2ef556557d7bd2bdd06d4eebe4296c7b929eee28b55"} Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.852097 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerStarted","Data":"ae7aead3b7c5a3e9931ebf74de8ed5ead9c4d5c4cf68dcd3f1e07a9eaa054b77"} Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.852135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerStarted","Data":"dacbee0bb3d865ee13bf64d6f1b7779c4f933c1c0aeb0c32e9b1e38138913715"} Dec 09 11:57:09 crc kubenswrapper[4770]: I1209 11:57:09.909280 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.099271 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.123967 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data\") pod \"72b4ab61-0de7-4073-986a-06cc75686d8a\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.124062 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tf5s\" (UniqueName: \"kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s\") pod \"72b4ab61-0de7-4073-986a-06cc75686d8a\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.124140 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle\") pod \"72b4ab61-0de7-4073-986a-06cc75686d8a\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.124181 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs\") pod \"72b4ab61-0de7-4073-986a-06cc75686d8a\" (UID: \"72b4ab61-0de7-4073-986a-06cc75686d8a\") " Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.125551 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs" (OuterVolumeSpecName: "logs") pod "72b4ab61-0de7-4073-986a-06cc75686d8a" (UID: "72b4ab61-0de7-4073-986a-06cc75686d8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.141266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s" (OuterVolumeSpecName: "kube-api-access-9tf5s") pod "72b4ab61-0de7-4073-986a-06cc75686d8a" (UID: "72b4ab61-0de7-4073-986a-06cc75686d8a"). InnerVolumeSpecName "kube-api-access-9tf5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.177515 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data" (OuterVolumeSpecName: "config-data") pod "72b4ab61-0de7-4073-986a-06cc75686d8a" (UID: "72b4ab61-0de7-4073-986a-06cc75686d8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.191014 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72b4ab61-0de7-4073-986a-06cc75686d8a" (UID: "72b4ab61-0de7-4073-986a-06cc75686d8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.229439 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tf5s\" (UniqueName: \"kubernetes.io/projected/72b4ab61-0de7-4073-986a-06cc75686d8a-kube-api-access-9tf5s\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.229487 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.229501 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72b4ab61-0de7-4073-986a-06cc75686d8a-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.229512 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b4ab61-0de7-4073-986a-06cc75686d8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.891310 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerStarted","Data":"5116de47af635a5b00f7da13b387beee7aab34ff8ea4ca434e4f6c2e3090bf91"} Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.897753 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.897752 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72b4ab61-0de7-4073-986a-06cc75686d8a","Type":"ContainerDied","Data":"5d49cd0f59e6f05376f2d7efab34348333f57b3939acd515cdae46de4aa46cbb"} Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.897923 4770 scope.go:117] "RemoveContainer" containerID="82557840fb376c018845c2ef556557d7bd2bdd06d4eebe4296c7b929eee28b55" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.899737 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3","Type":"ContainerStarted","Data":"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992"} Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.899788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3","Type":"ContainerStarted","Data":"4fa0d07083f52393eef891b1bfc559439e1d32adae87cdbebb79f801cbd7e6d6"} Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.925430 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.925405567 podStartE2EDuration="2.925405567s" podCreationTimestamp="2025-12-09 11:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:10.921835966 +0000 UTC m=+1556.162594485" watchObservedRunningTime="2025-12-09 11:57:10.925405567 +0000 UTC m=+1556.166164076" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.928128 4770 scope.go:117] "RemoveContainer" containerID="88cbccbcb372a92b554fc1b323246de00ed0703fe0f0670fef15cd010dc24750" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.978093 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.986274 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.995572 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:10 crc kubenswrapper[4770]: E1209 11:57:10.996130 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-api" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.996159 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-api" Dec 09 11:57:10 crc kubenswrapper[4770]: E1209 11:57:10.996201 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-log" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.996208 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-log" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.996405 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-api" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.996422 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" containerName="nova-api-log" Dec 09 11:57:10 crc kubenswrapper[4770]: I1209 11:57:10.997532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.000389 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.015986 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.049338 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.049406 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.049466 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hr4j\" (UniqueName: \"kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.049490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.058304 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.151158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hr4j\" (UniqueName: \"kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.151388 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.151598 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.151677 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.152315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.157457 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.172583 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.180670 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hr4j\" (UniqueName: \"kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j\") pod \"nova-api-0\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.355161 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b4ab61-0de7-4073-986a-06cc75686d8a" path="/var/lib/kubelet/pods/72b4ab61-0de7-4073-986a-06cc75686d8a/volumes" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.399302 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.897559 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:11 crc kubenswrapper[4770]: I1209 11:57:11.913785 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerStarted","Data":"7da7aaafcf09e00c4474192a91af6510a6ae9de2e1122d1925a532e82d5a71a3"} Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.925541 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerStarted","Data":"eaf666a59707d09ab24d8771ad533795627f89bf69142f938756699b674ac7a2"} Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.926504 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.928426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerStarted","Data":"997098b5337491d50b9c8750f8fc504bec865b43b538e12aa1546a5129387d0e"} Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.928549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerStarted","Data":"5c89d3d8f2b5884d22f703e41268c0b98dd4c95bab08249b005d5cd33f2510b5"} Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.928674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerStarted","Data":"3566741eb2638d64176acbfa53b1eecaed7402174d4e38764b74186424391f26"} Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.956689 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.460992592 podStartE2EDuration="4.956664004s" podCreationTimestamp="2025-12-09 11:57:08 +0000 UTC" firstStartedPulling="2025-12-09 11:57:08.911167471 +0000 UTC m=+1554.151925990" lastFinishedPulling="2025-12-09 11:57:12.406838883 +0000 UTC m=+1557.647597402" observedRunningTime="2025-12-09 11:57:12.947117362 +0000 UTC m=+1558.187875881" watchObservedRunningTime="2025-12-09 11:57:12.956664004 +0000 UTC m=+1558.197422523" Dec 09 11:57:12 crc kubenswrapper[4770]: I1209 11:57:12.974930 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9748879969999997 podStartE2EDuration="2.974887997s" podCreationTimestamp="2025-12-09 11:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:12.971125161 +0000 UTC m=+1558.211883690" watchObservedRunningTime="2025-12-09 11:57:12.974887997 +0000 UTC m=+1558.215646516" Dec 09 11:57:13 crc kubenswrapper[4770]: I1209 11:57:13.029579 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 11:57:14 crc kubenswrapper[4770]: I1209 11:57:14.277283 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.515819 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.518610 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.533537 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.535048 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.535342 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.535440 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rtm8\" (UniqueName: \"kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.637146 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.637288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.637323 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rtm8\" (UniqueName: \"kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.637672 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.637683 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.663104 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rtm8\" (UniqueName: \"kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8\") pod \"redhat-marketplace-fvhlv\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:18 crc kubenswrapper[4770]: I1209 11:57:18.860037 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:19 crc kubenswrapper[4770]: I1209 11:57:19.277449 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 11:57:19 crc kubenswrapper[4770]: I1209 11:57:19.312804 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 11:57:19 crc kubenswrapper[4770]: I1209 11:57:19.423216 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:19 crc kubenswrapper[4770]: I1209 11:57:19.999497 4770 generic.go:334] "Generic (PLEG): container finished" podID="4079db84-c9e9-493b-8828-48256a10d364" containerID="64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583" exitCode=0 Dec 09 11:57:19 crc kubenswrapper[4770]: I1209 11:57:19.999779 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerDied","Data":"64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583"} Dec 09 11:57:20 crc kubenswrapper[4770]: I1209 11:57:19.999867 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerStarted","Data":"62258823cf14604a277b671f1038da3d04c64c649f834b6dad08d5a47de88250"} Dec 09 11:57:20 crc kubenswrapper[4770]: I1209 11:57:20.032375 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 11:57:21 crc kubenswrapper[4770]: I1209 11:57:21.400519 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:57:21 crc kubenswrapper[4770]: I1209 11:57:21.400923 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:57:22 crc kubenswrapper[4770]: I1209 11:57:22.030098 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerStarted","Data":"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556"} Dec 09 11:57:22 crc kubenswrapper[4770]: I1209 11:57:22.484217 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:22 crc kubenswrapper[4770]: I1209 11:57:22.484345 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:23 crc kubenswrapper[4770]: I1209 11:57:23.042381 4770 generic.go:334] "Generic (PLEG): container finished" podID="4079db84-c9e9-493b-8828-48256a10d364" containerID="c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556" exitCode=0 Dec 09 11:57:23 crc kubenswrapper[4770]: I1209 11:57:23.042438 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerDied","Data":"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556"} Dec 09 11:57:24 crc kubenswrapper[4770]: I1209 11:57:24.157888 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerStarted","Data":"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726"} Dec 09 11:57:24 crc kubenswrapper[4770]: I1209 11:57:24.188443 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fvhlv" podStartSLOduration=2.745434596 podStartE2EDuration="6.18841431s" podCreationTimestamp="2025-12-09 11:57:18 +0000 UTC" firstStartedPulling="2025-12-09 11:57:20.002352618 +0000 UTC m=+1565.243111147" lastFinishedPulling="2025-12-09 11:57:23.445332342 +0000 UTC m=+1568.686090861" observedRunningTime="2025-12-09 11:57:24.18052114 +0000 UTC m=+1569.421279659" watchObservedRunningTime="2025-12-09 11:57:24.18841431 +0000 UTC m=+1569.429172829" Dec 09 11:57:26 crc kubenswrapper[4770]: I1209 11:57:26.971509 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:26 crc kubenswrapper[4770]: I1209 11:57:26.982592 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.000610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v5bt\" (UniqueName: \"kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt\") pod \"5e22b640-176b-4c5e-9747-190888df493a\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.001045 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data\") pod \"5e22b640-176b-4c5e-9747-190888df493a\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.001213 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle\") pod \"5e22b640-176b-4c5e-9747-190888df493a\" (UID: \"5e22b640-176b-4c5e-9747-190888df493a\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.020508 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt" (OuterVolumeSpecName: "kube-api-access-4v5bt") pod "5e22b640-176b-4c5e-9747-190888df493a" (UID: "5e22b640-176b-4c5e-9747-190888df493a"). InnerVolumeSpecName "kube-api-access-4v5bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.077464 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e22b640-176b-4c5e-9747-190888df493a" (UID: "5e22b640-176b-4c5e-9747-190888df493a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.077998 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data" (OuterVolumeSpecName: "config-data") pod "5e22b640-176b-4c5e-9747-190888df493a" (UID: "5e22b640-176b-4c5e-9747-190888df493a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.102417 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle\") pod \"84ade006-1109-4d4f-94f5-600457a7e178\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.102529 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data\") pod \"84ade006-1109-4d4f-94f5-600457a7e178\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.102627 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs\") pod \"84ade006-1109-4d4f-94f5-600457a7e178\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.102719 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw6h9\" (UniqueName: \"kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9\") pod \"84ade006-1109-4d4f-94f5-600457a7e178\" (UID: \"84ade006-1109-4d4f-94f5-600457a7e178\") " Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.103453 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v5bt\" (UniqueName: \"kubernetes.io/projected/5e22b640-176b-4c5e-9747-190888df493a-kube-api-access-4v5bt\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.103477 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.103491 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22b640-176b-4c5e-9747-190888df493a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.103584 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs" (OuterVolumeSpecName: "logs") pod "84ade006-1109-4d4f-94f5-600457a7e178" (UID: "84ade006-1109-4d4f-94f5-600457a7e178"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.106688 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9" (OuterVolumeSpecName: "kube-api-access-hw6h9") pod "84ade006-1109-4d4f-94f5-600457a7e178" (UID: "84ade006-1109-4d4f-94f5-600457a7e178"). InnerVolumeSpecName "kube-api-access-hw6h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.136293 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84ade006-1109-4d4f-94f5-600457a7e178" (UID: "84ade006-1109-4d4f-94f5-600457a7e178"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.137280 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data" (OuterVolumeSpecName: "config-data") pod "84ade006-1109-4d4f-94f5-600457a7e178" (UID: "84ade006-1109-4d4f-94f5-600457a7e178"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.188117 4770 generic.go:334] "Generic (PLEG): container finished" podID="5e22b640-176b-4c5e-9747-190888df493a" containerID="2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341" exitCode=137 Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.188479 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e22b640-176b-4c5e-9747-190888df493a","Type":"ContainerDied","Data":"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341"} Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.188665 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e22b640-176b-4c5e-9747-190888df493a","Type":"ContainerDied","Data":"40c241d9eedc435ff52f3efce46fa328324ebb3da3844a9c4e4408abf8000ee9"} Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.188764 4770 scope.go:117] "RemoveContainer" containerID="2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.188515 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.192365 4770 generic.go:334] "Generic (PLEG): container finished" podID="84ade006-1109-4d4f-94f5-600457a7e178" containerID="1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba" exitCode=137 Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.192529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerDied","Data":"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba"} Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.192665 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84ade006-1109-4d4f-94f5-600457a7e178","Type":"ContainerDied","Data":"8412495736e7a143e045411a2faab093f66ed3898e72b72af5d6ed7cc0e0a266"} Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.192834 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.205097 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.205129 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ade006-1109-4d4f-94f5-600457a7e178-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.205142 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84ade006-1109-4d4f-94f5-600457a7e178-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.205154 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw6h9\" (UniqueName: \"kubernetes.io/projected/84ade006-1109-4d4f-94f5-600457a7e178-kube-api-access-hw6h9\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.218986 4770 scope.go:117] "RemoveContainer" containerID="2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341" Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.222570 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341\": container with ID starting with 2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341 not found: ID does not exist" containerID="2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.222625 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341"} err="failed to get container status \"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341\": rpc error: code = NotFound desc = could not find container \"2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341\": container with ID starting with 2daa34308224f8b4ac73d4d10e99d0b5e9c7fa084be442a4e5dc70592d499341 not found: ID does not exist" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.222710 4770 scope.go:117] "RemoveContainer" containerID="1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.240555 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.267311 4770 scope.go:117] "RemoveContainer" containerID="6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.269589 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.284413 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.300119 4770 scope.go:117] "RemoveContainer" containerID="1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.300279 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.300752 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba\": container with ID starting with 1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba not found: ID does not exist" containerID="1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.300792 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba"} err="failed to get container status \"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba\": rpc error: code = NotFound desc = could not find container \"1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba\": container with ID starting with 1f29157af8b89103a0da10b813b4f05a427e81a934fc38ace1dc6cf10df133ba not found: ID does not exist" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.300826 4770 scope.go:117] "RemoveContainer" containerID="6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a" Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.301752 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a\": container with ID starting with 6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a not found: ID does not exist" containerID="6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.301786 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a"} err="failed to get container status \"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a\": rpc error: code = NotFound desc = could not find container \"6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a\": container with ID starting with 6313840d2e0cec47b2e1b868251ac54b972d10eb66e43bf688e31ae39c847c8a not found: ID does not exist" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.306404 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.307074 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-metadata" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307100 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-metadata" Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.307154 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e22b640-176b-4c5e-9747-190888df493a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307163 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e22b640-176b-4c5e-9747-190888df493a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:57:27 crc kubenswrapper[4770]: E1209 11:57:27.307191 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-log" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307200 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-log" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307446 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-log" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307493 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e22b640-176b-4c5e-9747-190888df493a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.307507 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ade006-1109-4d4f-94f5-600457a7e178" containerName="nova-metadata-metadata" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.308451 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.312032 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.312039 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.312084 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.319966 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.322471 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.331365 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.331538 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.333068 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.365067 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e22b640-176b-4c5e-9747-190888df493a" path="/var/lib/kubelet/pods/5e22b640-176b-4c5e-9747-190888df493a/volumes" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.365991 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84ade006-1109-4d4f-94f5-600457a7e178" path="/var/lib/kubelet/pods/84ade006-1109-4d4f-94f5-600457a7e178/volumes" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.366793 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.409883 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.409969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swdq8\" (UniqueName: \"kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.410051 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv79v\" (UniqueName: \"kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.410195 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.410405 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.411517 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.411557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.411653 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.411702 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.411852 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.513615 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.513708 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swdq8\" (UniqueName: \"kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.514291 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv79v\" (UniqueName: \"kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.514583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.514980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.515830 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.515888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.516007 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.516049 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.516218 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.516635 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.519492 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.519690 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.520026 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.520678 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.523990 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.524547 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.531716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.536302 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swdq8\" (UniqueName: \"kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8\") pod \"nova-metadata-0\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " pod="openstack/nova-metadata-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.536344 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv79v\" (UniqueName: \"kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.638343 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:27 crc kubenswrapper[4770]: I1209 11:57:27.650075 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.130379 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:28 crc kubenswrapper[4770]: W1209 11:57:28.131154 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod403c7fd8_7d16_4f89_89f7_f10ec283553c.slice/crio-4d6f10d7b0476fb623ef284a1c993e029638abd2f909dd2936b6afc6ea186465 WatchSource:0}: Error finding container 4d6f10d7b0476fb623ef284a1c993e029638abd2f909dd2936b6afc6ea186465: Status 404 returned error can't find the container with id 4d6f10d7b0476fb623ef284a1c993e029638abd2f909dd2936b6afc6ea186465 Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.210956 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerStarted","Data":"4d6f10d7b0476fb623ef284a1c993e029638abd2f909dd2936b6afc6ea186465"} Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.218360 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:57:28 crc kubenswrapper[4770]: W1209 11:57:28.222816 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6363bd5_219b_4fb9_9695_7924791e3262.slice/crio-d867aa5ccc5b0a777b2016533d93547869ee64ba952246a63fbcd7729967edc4 WatchSource:0}: Error finding container d867aa5ccc5b0a777b2016533d93547869ee64ba952246a63fbcd7729967edc4: Status 404 returned error can't find the container with id d867aa5ccc5b0a777b2016533d93547869ee64ba952246a63fbcd7729967edc4 Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.860696 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.861144 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:28 crc kubenswrapper[4770]: I1209 11:57:28.913888 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.228210 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerStarted","Data":"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b"} Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.229700 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerStarted","Data":"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7"} Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.233223 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6363bd5-219b-4fb9-9695-7924791e3262","Type":"ContainerStarted","Data":"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130"} Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.233285 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6363bd5-219b-4fb9-9695-7924791e3262","Type":"ContainerStarted","Data":"d867aa5ccc5b0a777b2016533d93547869ee64ba952246a63fbcd7729967edc4"} Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.268029 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.26800589 podStartE2EDuration="2.26800589s" podCreationTimestamp="2025-12-09 11:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:29.26091367 +0000 UTC m=+1574.501672229" watchObservedRunningTime="2025-12-09 11:57:29.26800589 +0000 UTC m=+1574.508764409" Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.301237 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.319459 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.319433416 podStartE2EDuration="2.319433416s" podCreationTimestamp="2025-12-09 11:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:29.282925649 +0000 UTC m=+1574.523684178" watchObservedRunningTime="2025-12-09 11:57:29.319433416 +0000 UTC m=+1574.560191935" Dec 09 11:57:29 crc kubenswrapper[4770]: I1209 11:57:29.369679 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.255954 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fvhlv" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="registry-server" containerID="cri-o://7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726" gracePeriod=2 Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.404666 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.406389 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.406925 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.409560 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.820441 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.917893 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content\") pod \"4079db84-c9e9-493b-8828-48256a10d364\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.918149 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities\") pod \"4079db84-c9e9-493b-8828-48256a10d364\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.918262 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rtm8\" (UniqueName: \"kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8\") pod \"4079db84-c9e9-493b-8828-48256a10d364\" (UID: \"4079db84-c9e9-493b-8828-48256a10d364\") " Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.919149 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities" (OuterVolumeSpecName: "utilities") pod "4079db84-c9e9-493b-8828-48256a10d364" (UID: "4079db84-c9e9-493b-8828-48256a10d364"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.924798 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8" (OuterVolumeSpecName: "kube-api-access-8rtm8") pod "4079db84-c9e9-493b-8828-48256a10d364" (UID: "4079db84-c9e9-493b-8828-48256a10d364"). InnerVolumeSpecName "kube-api-access-8rtm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:31 crc kubenswrapper[4770]: I1209 11:57:31.939993 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4079db84-c9e9-493b-8828-48256a10d364" (UID: "4079db84-c9e9-493b-8828-48256a10d364"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.021076 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.021130 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rtm8\" (UniqueName: \"kubernetes.io/projected/4079db84-c9e9-493b-8828-48256a10d364-kube-api-access-8rtm8\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.021145 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4079db84-c9e9-493b-8828-48256a10d364-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.269811 4770 generic.go:334] "Generic (PLEG): container finished" podID="4079db84-c9e9-493b-8828-48256a10d364" containerID="7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726" exitCode=0 Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.269873 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerDied","Data":"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726"} Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.269928 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvhlv" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.269972 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvhlv" event={"ID":"4079db84-c9e9-493b-8828-48256a10d364","Type":"ContainerDied","Data":"62258823cf14604a277b671f1038da3d04c64c649f834b6dad08d5a47de88250"} Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.270025 4770 scope.go:117] "RemoveContainer" containerID="7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.270671 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.277362 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.295287 4770 scope.go:117] "RemoveContainer" containerID="c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.328439 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.341843 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvhlv"] Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.344095 4770 scope.go:117] "RemoveContainer" containerID="64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.448661 4770 scope.go:117] "RemoveContainer" containerID="7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726" Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.456308 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726\": container with ID starting with 7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726 not found: ID does not exist" containerID="7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.456379 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726"} err="failed to get container status \"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726\": rpc error: code = NotFound desc = could not find container \"7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726\": container with ID starting with 7fda16b315d4b6530e6777c6a979c788bc868e77bef4ebbf13acfef5b65f7726 not found: ID does not exist" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.456423 4770 scope.go:117] "RemoveContainer" containerID="c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556" Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.460014 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556\": container with ID starting with c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556 not found: ID does not exist" containerID="c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.460075 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556"} err="failed to get container status \"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556\": rpc error: code = NotFound desc = could not find container \"c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556\": container with ID starting with c868d0f47bd32dbe465468a243317e20d0d63f7d093be3871c9cbc8bc53ae556 not found: ID does not exist" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.460114 4770 scope.go:117] "RemoveContainer" containerID="64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583" Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.463154 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583\": container with ID starting with 64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583 not found: ID does not exist" containerID="64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.463200 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583"} err="failed to get container status \"64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583\": rpc error: code = NotFound desc = could not find container \"64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583\": container with ID starting with 64cbd3ba2e4d5ea82ad6f6d182f1a1361cbde180e284787906448ede9f795583 not found: ID does not exist" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.475920 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.475994 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.481118 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.481573 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="extract-utilities" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.481586 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="extract-utilities" Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.481600 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="extract-content" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.481606 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="extract-content" Dec 09 11:57:32 crc kubenswrapper[4770]: E1209 11:57:32.481616 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="registry-server" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.481623 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="registry-server" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.481814 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4079db84-c9e9-493b-8828-48256a10d364" containerName="registry-server" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.482895 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.499268 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536495 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpqzv\" (UniqueName: \"kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536607 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536671 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536731 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.536830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.638583 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.638969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpqzv\" (UniqueName: \"kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.639117 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.639266 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.639398 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.639611 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.639845 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.640396 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.640599 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.640617 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.640757 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.641835 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.650460 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.650532 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.673278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpqzv\" (UniqueName: \"kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv\") pod \"dnsmasq-dns-867cd545c7-gzq6m\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.798979 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.813236 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.815124 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.825366 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.855333 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv7lz\" (UniqueName: \"kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.855431 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.855533 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.957791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv7lz\" (UniqueName: \"kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.958293 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.958349 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.959333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:32 crc kubenswrapper[4770]: I1209 11:57:32.959591 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:33 crc kubenswrapper[4770]: I1209 11:57:33.026287 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv7lz\" (UniqueName: \"kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz\") pod \"certified-operators-jj9qt\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:33 crc kubenswrapper[4770]: I1209 11:57:33.257218 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:33 crc kubenswrapper[4770]: I1209 11:57:33.355659 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4079db84-c9e9-493b-8828-48256a10d364" path="/var/lib/kubelet/pods/4079db84-c9e9-493b-8828-48256a10d364/volumes" Dec 09 11:57:33 crc kubenswrapper[4770]: I1209 11:57:33.358059 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:57:33 crc kubenswrapper[4770]: W1209 11:57:33.359937 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde2cc3bf_23a6_4b3f_a1fa_b845aad71787.slice/crio-73989f64741448fd7cfbe906449b5e2863fedf1adec7d51f6605c5762a8d2648 WatchSource:0}: Error finding container 73989f64741448fd7cfbe906449b5e2863fedf1adec7d51f6605c5762a8d2648: Status 404 returned error can't find the container with id 73989f64741448fd7cfbe906449b5e2863fedf1adec7d51f6605c5762a8d2648 Dec 09 11:57:33 crc kubenswrapper[4770]: I1209 11:57:33.801089 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:33 crc kubenswrapper[4770]: W1209 11:57:33.837844 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab99f8c_4cdf_44b0_a4ba_9e243570aaa8.slice/crio-2754f1fafc39960bf64c7158d13e8be36fb3f0fce2fd50c3b17f422ec6589202 WatchSource:0}: Error finding container 2754f1fafc39960bf64c7158d13e8be36fb3f0fce2fd50c3b17f422ec6589202: Status 404 returned error can't find the container with id 2754f1fafc39960bf64c7158d13e8be36fb3f0fce2fd50c3b17f422ec6589202 Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.294763 4770 generic.go:334] "Generic (PLEG): container finished" podID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerID="06e985e24579ff9c413e840e9b0638c771aa22054a9c57c3468b29de99f1d1da" exitCode=0 Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.294847 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" event={"ID":"de2cc3bf-23a6-4b3f-a1fa-b845aad71787","Type":"ContainerDied","Data":"06e985e24579ff9c413e840e9b0638c771aa22054a9c57c3468b29de99f1d1da"} Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.295140 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" event={"ID":"de2cc3bf-23a6-4b3f-a1fa-b845aad71787","Type":"ContainerStarted","Data":"73989f64741448fd7cfbe906449b5e2863fedf1adec7d51f6605c5762a8d2648"} Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.298731 4770 generic.go:334] "Generic (PLEG): container finished" podID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerID="8ac4466d5178f15464c2e1982e8f6878241fbedc6515386a2a730607eac26d12" exitCode=0 Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.298809 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerDied","Data":"8ac4466d5178f15464c2e1982e8f6878241fbedc6515386a2a730607eac26d12"} Dec 09 11:57:34 crc kubenswrapper[4770]: I1209 11:57:34.298878 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerStarted","Data":"2754f1fafc39960bf64c7158d13e8be36fb3f0fce2fd50c3b17f422ec6589202"} Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.313320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" event={"ID":"de2cc3bf-23a6-4b3f-a1fa-b845aad71787","Type":"ContainerStarted","Data":"43ebc26e2c2fa580b48c4844f5bfe5b3ba67d22ccd82e4aaaa68755ef5a9c123"} Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.313655 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.317332 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerStarted","Data":"1e40cdbf3f62f722d2c55dd654186a03cfc4bf79cb93901d6d18c476fd45ace6"} Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.336226 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" podStartSLOduration=3.336176324 podStartE2EDuration="3.336176324s" podCreationTimestamp="2025-12-09 11:57:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:35.335977569 +0000 UTC m=+1580.576736108" watchObservedRunningTime="2025-12-09 11:57:35.336176324 +0000 UTC m=+1580.576934853" Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.549867 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.550577 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-central-agent" containerID="cri-o://ae7aead3b7c5a3e9931ebf74de8ed5ead9c4d5c4cf68dcd3f1e07a9eaa054b77" gracePeriod=30 Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.550774 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" containerID="cri-o://eaf666a59707d09ab24d8771ad533795627f89bf69142f938756699b674ac7a2" gracePeriod=30 Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.550835 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="sg-core" containerID="cri-o://7da7aaafcf09e00c4474192a91af6510a6ae9de2e1122d1925a532e82d5a71a3" gracePeriod=30 Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.550954 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-notification-agent" containerID="cri-o://5116de47af635a5b00f7da13b387beee7aab34ff8ea4ca434e4f6c2e3090bf91" gracePeriod=30 Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.573826 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.739229 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.740236 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-log" containerID="cri-o://5c89d3d8f2b5884d22f703e41268c0b98dd4c95bab08249b005d5cd33f2510b5" gracePeriod=30 Dec 09 11:57:35 crc kubenswrapper[4770]: I1209 11:57:35.740344 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-api" containerID="cri-o://997098b5337491d50b9c8750f8fc504bec865b43b538e12aa1546a5129387d0e" gracePeriod=30 Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.342329 4770 generic.go:334] "Generic (PLEG): container finished" podID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerID="eaf666a59707d09ab24d8771ad533795627f89bf69142f938756699b674ac7a2" exitCode=0 Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.342372 4770 generic.go:334] "Generic (PLEG): container finished" podID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerID="7da7aaafcf09e00c4474192a91af6510a6ae9de2e1122d1925a532e82d5a71a3" exitCode=2 Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.342423 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerDied","Data":"eaf666a59707d09ab24d8771ad533795627f89bf69142f938756699b674ac7a2"} Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.342451 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerDied","Data":"7da7aaafcf09e00c4474192a91af6510a6ae9de2e1122d1925a532e82d5a71a3"} Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.354027 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d23d4b3-f359-4da0-975b-39459c100048" containerID="5c89d3d8f2b5884d22f703e41268c0b98dd4c95bab08249b005d5cd33f2510b5" exitCode=143 Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.354404 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerDied","Data":"5c89d3d8f2b5884d22f703e41268c0b98dd4c95bab08249b005d5cd33f2510b5"} Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.374659 4770 generic.go:334] "Generic (PLEG): container finished" podID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerID="1e40cdbf3f62f722d2c55dd654186a03cfc4bf79cb93901d6d18c476fd45ace6" exitCode=0 Dec 09 11:57:36 crc kubenswrapper[4770]: I1209 11:57:36.376061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerDied","Data":"1e40cdbf3f62f722d2c55dd654186a03cfc4bf79cb93901d6d18c476fd45ace6"} Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.389684 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerStarted","Data":"bf8461b595d392e1ff8c8bc01bac840993df317a04c41965ab4d403296f8c44f"} Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.397612 4770 generic.go:334] "Generic (PLEG): container finished" podID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerID="ae7aead3b7c5a3e9931ebf74de8ed5ead9c4d5c4cf68dcd3f1e07a9eaa054b77" exitCode=0 Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.397684 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerDied","Data":"ae7aead3b7c5a3e9931ebf74de8ed5ead9c4d5c4cf68dcd3f1e07a9eaa054b77"} Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.418057 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jj9qt" podStartSLOduration=2.864933148 podStartE2EDuration="5.418035156s" podCreationTimestamp="2025-12-09 11:57:32 +0000 UTC" firstStartedPulling="2025-12-09 11:57:34.30066984 +0000 UTC m=+1579.541428359" lastFinishedPulling="2025-12-09 11:57:36.853771848 +0000 UTC m=+1582.094530367" observedRunningTime="2025-12-09 11:57:37.4150401 +0000 UTC m=+1582.655798629" watchObservedRunningTime="2025-12-09 11:57:37.418035156 +0000 UTC m=+1582.658793675" Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.638647 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.650798 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.651029 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 11:57:37 crc kubenswrapper[4770]: I1209 11:57:37.679042 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.417502 4770 generic.go:334] "Generic (PLEG): container finished" podID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerID="5116de47af635a5b00f7da13b387beee7aab34ff8ea4ca434e4f6c2e3090bf91" exitCode=0 Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.417704 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerDied","Data":"5116de47af635a5b00f7da13b387beee7aab34ff8ea4ca434e4f6c2e3090bf91"} Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.418121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83f8d015-ef90-4b96-ad91-8c6ffa850a25","Type":"ContainerDied","Data":"dacbee0bb3d865ee13bf64d6f1b7779c4f933c1c0aeb0c32e9b1e38138913715"} Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.418152 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dacbee0bb3d865ee13bf64d6f1b7779c4f933c1c0aeb0c32e9b1e38138913715" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.439365 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.533791 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593320 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593370 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593411 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593438 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrj6c\" (UniqueName: \"kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593479 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.593526 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.594510 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.615958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts" (OuterVolumeSpecName: "scripts") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.620221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c" (OuterVolumeSpecName: "kube-api-access-jrj6c") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "kube-api-access-jrj6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.650087 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.665477 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.665939 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.677412 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-qzmrd"] Dec 09 11:57:38 crc kubenswrapper[4770]: E1209 11:57:38.679230 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-notification-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679262 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-notification-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: E1209 11:57:38.679279 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-central-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679289 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-central-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: E1209 11:57:38.679330 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="sg-core" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679339 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="sg-core" Dec 09 11:57:38 crc kubenswrapper[4770]: E1209 11:57:38.679364 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679373 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679621 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679645 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-central-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679657 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="sg-core" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.679672 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="ceilometer-notification-agent" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.680631 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.686123 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.704051 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.713709 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.739927 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qzmrd"] Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.741398 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.741476 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data\") pod \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\" (UID: \"83f8d015-ef90-4b96-ad91-8c6ffa850a25\") " Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.743775 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.742304 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.744526 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftj7v\" (UniqueName: \"kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.744792 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.744821 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745209 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745238 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745252 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83f8d015-ef90-4b96-ad91-8c6ffa850a25-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745264 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745284 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.745295 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrj6c\" (UniqueName: \"kubernetes.io/projected/83f8d015-ef90-4b96-ad91-8c6ffa850a25-kube-api-access-jrj6c\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.788398 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.847172 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.847294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftj7v\" (UniqueName: \"kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.847427 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.847452 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.847541 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.852384 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.852381 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.853958 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.866706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftj7v\" (UniqueName: \"kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v\") pod \"nova-cell1-cell-mapping-qzmrd\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.882866 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data" (OuterVolumeSpecName: "config-data") pod "83f8d015-ef90-4b96-ad91-8c6ffa850a25" (UID: "83f8d015-ef90-4b96-ad91-8c6ffa850a25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:38 crc kubenswrapper[4770]: I1209 11:57:38.949374 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f8d015-ef90-4b96-ad91-8c6ffa850a25-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.072350 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.443196 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d23d4b3-f359-4da0-975b-39459c100048" containerID="997098b5337491d50b9c8750f8fc504bec865b43b538e12aa1546a5129387d0e" exitCode=0 Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.443269 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerDied","Data":"997098b5337491d50b9c8750f8fc504bec865b43b538e12aa1546a5129387d0e"} Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.443736 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.524346 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.538810 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.548062 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.550385 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.553179 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.553624 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.553819 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.563707 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.629025 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qzmrd"] Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.685953 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9xkt\" (UniqueName: \"kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686340 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686402 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686458 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686474 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686529 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.686583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.720480 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789094 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data\") pod \"0d23d4b3-f359-4da0-975b-39459c100048\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789217 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs\") pod \"0d23d4b3-f359-4da0-975b-39459c100048\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hr4j\" (UniqueName: \"kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j\") pod \"0d23d4b3-f359-4da0-975b-39459c100048\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789397 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle\") pod \"0d23d4b3-f359-4da0-975b-39459c100048\" (UID: \"0d23d4b3-f359-4da0-975b-39459c100048\") " Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789652 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789681 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789704 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789738 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789787 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789830 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9xkt\" (UniqueName: \"kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.789879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.790052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.791239 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.791929 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs" (OuterVolumeSpecName: "logs") pod "0d23d4b3-f359-4da0-975b-39459c100048" (UID: "0d23d4b3-f359-4da0-975b-39459c100048"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.793086 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.796707 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.799238 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.800269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.805751 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.806460 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.811425 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j" (OuterVolumeSpecName: "kube-api-access-7hr4j") pod "0d23d4b3-f359-4da0-975b-39459c100048" (UID: "0d23d4b3-f359-4da0-975b-39459c100048"). InnerVolumeSpecName "kube-api-access-7hr4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.835854 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9xkt\" (UniqueName: \"kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt\") pod \"ceilometer-0\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.838117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d23d4b3-f359-4da0-975b-39459c100048" (UID: "0d23d4b3-f359-4da0-975b-39459c100048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.838740 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data" (OuterVolumeSpecName: "config-data") pod "0d23d4b3-f359-4da0-975b-39459c100048" (UID: "0d23d4b3-f359-4da0-975b-39459c100048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.890987 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.891806 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d23d4b3-f359-4da0-975b-39459c100048-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.891838 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hr4j\" (UniqueName: \"kubernetes.io/projected/0d23d4b3-f359-4da0-975b-39459c100048-kube-api-access-7hr4j\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.891852 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:39 crc kubenswrapper[4770]: I1209 11:57:39.891861 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23d4b3-f359-4da0-975b-39459c100048-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:40 crc kubenswrapper[4770]: W1209 11:57:40.370846 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7e5fb15_de40_44d2_8e1c_fcb78a080d38.slice/crio-6950ccae8d1bffad99d338e0ae384434148fe8c6b289dce6029995ac77c6226a WatchSource:0}: Error finding container 6950ccae8d1bffad99d338e0ae384434148fe8c6b289dce6029995ac77c6226a: Status 404 returned error can't find the container with id 6950ccae8d1bffad99d338e0ae384434148fe8c6b289dce6029995ac77c6226a Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.372088 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.462108 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerStarted","Data":"6950ccae8d1bffad99d338e0ae384434148fe8c6b289dce6029995ac77c6226a"} Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.465486 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qzmrd" event={"ID":"a8ce6a43-137d-49d2-8f22-d72d01e55ae5","Type":"ContainerStarted","Data":"8e69aba5ad2edf194e1fbfdc44ca9cb279f74fad72141a9489720b1480d0b44c"} Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.465534 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qzmrd" event={"ID":"a8ce6a43-137d-49d2-8f22-d72d01e55ae5","Type":"ContainerStarted","Data":"fcd47da709079d77e6ea85f78b907bd6fc07461b88a4b722008667b97831e6d3"} Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.469833 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d23d4b3-f359-4da0-975b-39459c100048","Type":"ContainerDied","Data":"3566741eb2638d64176acbfa53b1eecaed7402174d4e38764b74186424391f26"} Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.469892 4770 scope.go:117] "RemoveContainer" containerID="997098b5337491d50b9c8750f8fc504bec865b43b538e12aa1546a5129387d0e" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.470106 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.507090 4770 scope.go:117] "RemoveContainer" containerID="5c89d3d8f2b5884d22f703e41268c0b98dd4c95bab08249b005d5cd33f2510b5" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.513764 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-qzmrd" podStartSLOduration=2.513738451 podStartE2EDuration="2.513738451s" podCreationTimestamp="2025-12-09 11:57:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:40.495673083 +0000 UTC m=+1585.736431612" watchObservedRunningTime="2025-12-09 11:57:40.513738451 +0000 UTC m=+1585.754496961" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.535997 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.549425 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.577824 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:40 crc kubenswrapper[4770]: E1209 11:57:40.578411 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-api" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.578436 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-api" Dec 09 11:57:40 crc kubenswrapper[4770]: E1209 11:57:40.578460 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-log" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.578468 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-log" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.578662 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-api" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.578686 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d23d4b3-f359-4da0-975b-39459c100048" containerName="nova-api-log" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.579893 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.584662 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.584734 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.584673 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.599429 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.713788 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.713932 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.713985 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.714016 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.714052 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87ldc\" (UniqueName: \"kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.714097 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816192 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816279 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87ldc\" (UniqueName: \"kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816356 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816409 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816529 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.816592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.818184 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.824388 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.824528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.826076 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.826493 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.837889 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87ldc\" (UniqueName: \"kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc\") pod \"nova-api-0\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " pod="openstack/nova-api-0" Dec 09 11:57:40 crc kubenswrapper[4770]: I1209 11:57:40.907756 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:41 crc kubenswrapper[4770]: I1209 11:57:41.386038 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d23d4b3-f359-4da0-975b-39459c100048" path="/var/lib/kubelet/pods/0d23d4b3-f359-4da0-975b-39459c100048/volumes" Dec 09 11:57:41 crc kubenswrapper[4770]: I1209 11:57:41.387602 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" path="/var/lib/kubelet/pods/83f8d015-ef90-4b96-ad91-8c6ffa850a25/volumes" Dec 09 11:57:41 crc kubenswrapper[4770]: I1209 11:57:41.427046 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:41 crc kubenswrapper[4770]: I1209 11:57:41.503259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerStarted","Data":"e377a23d626f32c3c0b13a20c65b02b97f6a837c0852508e8b08876244ab3726"} Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.523407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerStarted","Data":"07ea87393e7d08e9bbfe345d3786c9fd9565d8d1f4171600282eccd2c3cc50f2"} Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.527574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerStarted","Data":"e61d75bbd84a9c596e11b0fc0fc19cffb2909291bdc0df0a77fcc6ca26164855"} Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.527629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerStarted","Data":"2da012ee64e95c1c9a9324467e7a92feaccf4db547813cfe6304166e23cd0dad"} Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.554842 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.554811869 podStartE2EDuration="2.554811869s" podCreationTimestamp="2025-12-09 11:57:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:42.548069308 +0000 UTC m=+1587.788827827" watchObservedRunningTime="2025-12-09 11:57:42.554811869 +0000 UTC m=+1587.795570398" Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.819128 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.926539 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:57:42 crc kubenswrapper[4770]: I1209 11:57:42.927196 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="dnsmasq-dns" containerID="cri-o://e143040a64c122594f975a4e26cc8fc4b0b71048af4090cbf1c907cc2b1720dc" gracePeriod=10 Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.258567 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.259847 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.338310 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.543694 4770 generic.go:334] "Generic (PLEG): container finished" podID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerID="e143040a64c122594f975a4e26cc8fc4b0b71048af4090cbf1c907cc2b1720dc" exitCode=0 Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.543743 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" event={"ID":"07fb6c2a-e734-4169-9f47-c093a215f9ec","Type":"ContainerDied","Data":"e143040a64c122594f975a4e26cc8fc4b0b71048af4090cbf1c907cc2b1720dc"} Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.547127 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerStarted","Data":"bd1a0426233ccd1b1386234cfd700d23db549b81974daa36103a0f808577669b"} Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.611522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.677260 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.843443 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.885951 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.886026 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwl65\" (UniqueName: \"kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.886118 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.886171 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.886210 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.886313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb\") pod \"07fb6c2a-e734-4169-9f47-c093a215f9ec\" (UID: \"07fb6c2a-e734-4169-9f47-c093a215f9ec\") " Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.894385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65" (OuterVolumeSpecName: "kube-api-access-gwl65") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "kube-api-access-gwl65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.973582 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.983175 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.989693 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwl65\" (UniqueName: \"kubernetes.io/projected/07fb6c2a-e734-4169-9f47-c093a215f9ec-kube-api-access-gwl65\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.989724 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.989735 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.993075 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:43 crc kubenswrapper[4770]: I1209 11:57:43.998351 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config" (OuterVolumeSpecName: "config") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.006398 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "07fb6c2a-e734-4169-9f47-c093a215f9ec" (UID: "07fb6c2a-e734-4169-9f47-c093a215f9ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.092114 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.092182 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.092194 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07fb6c2a-e734-4169-9f47-c093a215f9ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.558573 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" event={"ID":"07fb6c2a-e734-4169-9f47-c093a215f9ec","Type":"ContainerDied","Data":"9416370cef5624f158deb7862fc13e2e987801a42abf44957346b6d155484d22"} Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.558628 4770 scope.go:117] "RemoveContainer" containerID="e143040a64c122594f975a4e26cc8fc4b0b71048af4090cbf1c907cc2b1720dc" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.558645 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb54f9b5-tk688" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.562694 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerStarted","Data":"b85b05cd17651d9b3bd353aa522c2eeb740da829a60de5831bcd5ea98ce4117d"} Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.590460 4770 scope.go:117] "RemoveContainer" containerID="73acd5b0d78a865d73b43f88829aa4a826410576bc0db186ae87a7c400450bbb" Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.606846 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:57:44 crc kubenswrapper[4770]: I1209 11:57:44.620398 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bfb54f9b5-tk688"] Dec 09 11:57:45 crc kubenswrapper[4770]: I1209 11:57:45.353390 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" path="/var/lib/kubelet/pods/07fb6c2a-e734-4169-9f47-c093a215f9ec/volumes" Dec 09 11:57:45 crc kubenswrapper[4770]: I1209 11:57:45.574778 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jj9qt" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="registry-server" containerID="cri-o://bf8461b595d392e1ff8c8bc01bac840993df317a04c41965ab4d403296f8c44f" gracePeriod=2 Dec 09 11:57:46 crc kubenswrapper[4770]: I1209 11:57:46.588541 4770 generic.go:334] "Generic (PLEG): container finished" podID="a8ce6a43-137d-49d2-8f22-d72d01e55ae5" containerID="8e69aba5ad2edf194e1fbfdc44ca9cb279f74fad72141a9489720b1480d0b44c" exitCode=0 Dec 09 11:57:46 crc kubenswrapper[4770]: I1209 11:57:46.588622 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qzmrd" event={"ID":"a8ce6a43-137d-49d2-8f22-d72d01e55ae5","Type":"ContainerDied","Data":"8e69aba5ad2edf194e1fbfdc44ca9cb279f74fad72141a9489720b1480d0b44c"} Dec 09 11:57:46 crc kubenswrapper[4770]: I1209 11:57:46.593163 4770 generic.go:334] "Generic (PLEG): container finished" podID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerID="bf8461b595d392e1ff8c8bc01bac840993df317a04c41965ab4d403296f8c44f" exitCode=0 Dec 09 11:57:46 crc kubenswrapper[4770]: I1209 11:57:46.593234 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerDied","Data":"bf8461b595d392e1ff8c8bc01bac840993df317a04c41965ab4d403296f8c44f"} Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.304427 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.361935 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv7lz\" (UniqueName: \"kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz\") pod \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.362545 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities\") pod \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.362670 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content\") pod \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\" (UID: \"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8\") " Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.365171 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities" (OuterVolumeSpecName: "utilities") pod "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" (UID: "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.371481 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz" (OuterVolumeSpecName: "kube-api-access-cv7lz") pod "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" (UID: "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8"). InnerVolumeSpecName "kube-api-access-cv7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.427864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" (UID: "eab99f8c-4cdf-44b0-a4ba-9e243570aaa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.466653 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.466714 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.466732 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv7lz\" (UniqueName: \"kubernetes.io/projected/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8-kube-api-access-cv7lz\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.606304 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9qt" event={"ID":"eab99f8c-4cdf-44b0-a4ba-9e243570aaa8","Type":"ContainerDied","Data":"2754f1fafc39960bf64c7158d13e8be36fb3f0fce2fd50c3b17f422ec6589202"} Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.606368 4770 scope.go:117] "RemoveContainer" containerID="bf8461b595d392e1ff8c8bc01bac840993df317a04c41965ab4d403296f8c44f" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.606522 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9qt" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.619840 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerStarted","Data":"d3cff712577cb5ca4995f819406108099e3cabf583e0b9fe83d7a120676f4548"} Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.620121 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.656386 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.491159399 podStartE2EDuration="8.656357326s" podCreationTimestamp="2025-12-09 11:57:39 +0000 UTC" firstStartedPulling="2025-12-09 11:57:40.375475821 +0000 UTC m=+1585.616234340" lastFinishedPulling="2025-12-09 11:57:46.540673748 +0000 UTC m=+1591.781432267" observedRunningTime="2025-12-09 11:57:47.647121522 +0000 UTC m=+1592.887880051" watchObservedRunningTime="2025-12-09 11:57:47.656357326 +0000 UTC m=+1592.897115845" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.665013 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.665763 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.670997 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.678918 4770 scope.go:117] "RemoveContainer" containerID="1e40cdbf3f62f722d2c55dd654186a03cfc4bf79cb93901d6d18c476fd45ace6" Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.682054 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.698448 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jj9qt"] Dec 09 11:57:47 crc kubenswrapper[4770]: I1209 11:57:47.724746 4770 scope.go:117] "RemoveContainer" containerID="8ac4466d5178f15464c2e1982e8f6878241fbedc6515386a2a730607eac26d12" Dec 09 11:57:47 crc kubenswrapper[4770]: E1209 11:57:47.788846 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab99f8c_4cdf_44b0_a4ba_9e243570aaa8.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.006102 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.083999 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftj7v\" (UniqueName: \"kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v\") pod \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.084095 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle\") pod \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.084237 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts\") pod \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.084287 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data\") pod \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\" (UID: \"a8ce6a43-137d-49d2-8f22-d72d01e55ae5\") " Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.090624 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts" (OuterVolumeSpecName: "scripts") pod "a8ce6a43-137d-49d2-8f22-d72d01e55ae5" (UID: "a8ce6a43-137d-49d2-8f22-d72d01e55ae5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.090921 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v" (OuterVolumeSpecName: "kube-api-access-ftj7v") pod "a8ce6a43-137d-49d2-8f22-d72d01e55ae5" (UID: "a8ce6a43-137d-49d2-8f22-d72d01e55ae5"). InnerVolumeSpecName "kube-api-access-ftj7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.117168 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data" (OuterVolumeSpecName: "config-data") pod "a8ce6a43-137d-49d2-8f22-d72d01e55ae5" (UID: "a8ce6a43-137d-49d2-8f22-d72d01e55ae5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.119097 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8ce6a43-137d-49d2-8f22-d72d01e55ae5" (UID: "a8ce6a43-137d-49d2-8f22-d72d01e55ae5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.187141 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftj7v\" (UniqueName: \"kubernetes.io/projected/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-kube-api-access-ftj7v\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.187689 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.187701 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.187710 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce6a43-137d-49d2-8f22-d72d01e55ae5-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.643377 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qzmrd" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.644340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qzmrd" event={"ID":"a8ce6a43-137d-49d2-8f22-d72d01e55ae5","Type":"ContainerDied","Data":"fcd47da709079d77e6ea85f78b907bd6fc07461b88a4b722008667b97831e6d3"} Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.644375 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcd47da709079d77e6ea85f78b907bd6fc07461b88a4b722008667b97831e6d3" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.653597 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.811750 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.812175 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-log" containerID="cri-o://2da012ee64e95c1c9a9324467e7a92feaccf4db547813cfe6304166e23cd0dad" gracePeriod=30 Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.812282 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-api" containerID="cri-o://e61d75bbd84a9c596e11b0fc0fc19cffb2909291bdc0df0a77fcc6ca26164855" gracePeriod=30 Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.828008 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.828348 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" containerID="cri-o://dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" gracePeriod=30 Dec 09 11:57:48 crc kubenswrapper[4770]: I1209 11:57:48.850424 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:49 crc kubenswrapper[4770]: E1209 11:57:49.279657 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:49 crc kubenswrapper[4770]: E1209 11:57:49.281688 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:49 crc kubenswrapper[4770]: E1209 11:57:49.282721 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:49 crc kubenswrapper[4770]: E1209 11:57:49.282789 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.353843 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" path="/var/lib/kubelet/pods/eab99f8c-4cdf-44b0-a4ba-9e243570aaa8/volumes" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.704670 4770 generic.go:334] "Generic (PLEG): container finished" podID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerID="e61d75bbd84a9c596e11b0fc0fc19cffb2909291bdc0df0a77fcc6ca26164855" exitCode=0 Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.705002 4770 generic.go:334] "Generic (PLEG): container finished" podID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerID="2da012ee64e95c1c9a9324467e7a92feaccf4db547813cfe6304166e23cd0dad" exitCode=143 Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.704730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerDied","Data":"e61d75bbd84a9c596e11b0fc0fc19cffb2909291bdc0df0a77fcc6ca26164855"} Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.705217 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerDied","Data":"2da012ee64e95c1c9a9324467e7a92feaccf4db547813cfe6304166e23cd0dad"} Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.879086 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919167 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919296 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919345 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919387 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919405 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87ldc\" (UniqueName: \"kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.919425 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data\") pod \"ea541475-b6fc-464e-b0c2-f347aeb91d64\" (UID: \"ea541475-b6fc-464e-b0c2-f347aeb91d64\") " Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.920683 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs" (OuterVolumeSpecName: "logs") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.928390 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc" (OuterVolumeSpecName: "kube-api-access-87ldc") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "kube-api-access-87ldc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.957470 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data" (OuterVolumeSpecName: "config-data") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.958221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.984586 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:49 crc kubenswrapper[4770]: I1209 11:57:49.998916 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ea541475-b6fc-464e-b0c2-f347aeb91d64" (UID: "ea541475-b6fc-464e-b0c2-f347aeb91d64"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021653 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021683 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea541475-b6fc-464e-b0c2-f347aeb91d64-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021694 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021703 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87ldc\" (UniqueName: \"kubernetes.io/projected/ea541475-b6fc-464e-b0c2-f347aeb91d64-kube-api-access-87ldc\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021716 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.021727 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea541475-b6fc-464e-b0c2-f347aeb91d64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.716886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea541475-b6fc-464e-b0c2-f347aeb91d64","Type":"ContainerDied","Data":"e377a23d626f32c3c0b13a20c65b02b97f6a837c0852508e8b08876244ab3726"} Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.717244 4770 scope.go:117] "RemoveContainer" containerID="e61d75bbd84a9c596e11b0fc0fc19cffb2909291bdc0df0a77fcc6ca26164855" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.716950 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.717137 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" containerID="cri-o://e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b" gracePeriod=30 Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.717111 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" containerID="cri-o://0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7" gracePeriod=30 Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.749090 4770 scope.go:117] "RemoveContainer" containerID="2da012ee64e95c1c9a9324467e7a92feaccf4db547813cfe6304166e23cd0dad" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.759384 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.773666 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.785876 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786446 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="extract-content" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786465 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="extract-content" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786483 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ce6a43-137d-49d2-8f22-d72d01e55ae5" containerName="nova-manage" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786490 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ce6a43-137d-49d2-8f22-d72d01e55ae5" containerName="nova-manage" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786502 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="dnsmasq-dns" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786508 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="dnsmasq-dns" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786517 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-api" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786523 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-api" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786538 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-log" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786544 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-log" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786561 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="init" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786567 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="init" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786580 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="extract-utilities" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786588 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="extract-utilities" Dec 09 11:57:50 crc kubenswrapper[4770]: E1209 11:57:50.786600 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="registry-server" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786606 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="registry-server" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786786 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-log" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786800 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ce6a43-137d-49d2-8f22-d72d01e55ae5" containerName="nova-manage" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786818 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="07fb6c2a-e734-4169-9f47-c093a215f9ec" containerName="dnsmasq-dns" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786833 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab99f8c-4cdf-44b0-a4ba-9e243570aaa8" containerName="registry-server" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.786842 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" containerName="nova-api-api" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.788073 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.790276 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.792280 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.792336 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.798002 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.880660 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.880885 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.881023 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp6jk\" (UniqueName: \"kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.881043 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.881114 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.881151 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982500 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982619 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp6jk\" (UniqueName: \"kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982652 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982727 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.982773 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.983037 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.988417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.988467 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.988462 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:50 crc kubenswrapper[4770]: I1209 11:57:50.990658 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.006549 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp6jk\" (UniqueName: \"kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk\") pod \"nova-api-0\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " pod="openstack/nova-api-0" Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.154430 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.355981 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea541475-b6fc-464e-b0c2-f347aeb91d64" path="/var/lib/kubelet/pods/ea541475-b6fc-464e-b0c2-f347aeb91d64/volumes" Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.647848 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:57:51 crc kubenswrapper[4770]: W1209 11:57:51.648871 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc923d25_3beb_45cd_a504_0ebb6391a9f4.slice/crio-7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8 WatchSource:0}: Error finding container 7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8: Status 404 returned error can't find the container with id 7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8 Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.733750 4770 generic.go:334] "Generic (PLEG): container finished" podID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerID="0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7" exitCode=143 Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.733872 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerDied","Data":"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7"} Dec 09 11:57:51 crc kubenswrapper[4770]: I1209 11:57:51.735435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerStarted","Data":"7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8"} Dec 09 11:57:52 crc kubenswrapper[4770]: I1209 11:57:52.749682 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerStarted","Data":"9a473cbcc51156e661f62a9cb155beb880dd96bd5e5e0ee00faf14c8c0beeb93"} Dec 09 11:57:52 crc kubenswrapper[4770]: I1209 11:57:52.750061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerStarted","Data":"243be21689d2a29c2966994727f43b6692216066e84a9a1e11711c1c2553ffb8"} Dec 09 11:57:53 crc kubenswrapper[4770]: I1209 11:57:53.855263 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:58804->10.217.0.194:8775: read: connection reset by peer" Dec 09 11:57:53 crc kubenswrapper[4770]: I1209 11:57:53.855819 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:58800->10.217.0.194:8775: read: connection reset by peer" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.288296 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.290496 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.292819 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.292873 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.335914 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.365334 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.36531068 podStartE2EDuration="4.36531068s" podCreationTimestamp="2025-12-09 11:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:52.781617387 +0000 UTC m=+1598.022375906" watchObservedRunningTime="2025-12-09 11:57:54.36531068 +0000 UTC m=+1599.606069199" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.470692 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs\") pod \"403c7fd8-7d16-4f89-89f7-f10ec283553c\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.470795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data\") pod \"403c7fd8-7d16-4f89-89f7-f10ec283553c\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.470922 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swdq8\" (UniqueName: \"kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8\") pod \"403c7fd8-7d16-4f89-89f7-f10ec283553c\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.471169 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs\") pod \"403c7fd8-7d16-4f89-89f7-f10ec283553c\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.471247 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle\") pod \"403c7fd8-7d16-4f89-89f7-f10ec283553c\" (UID: \"403c7fd8-7d16-4f89-89f7-f10ec283553c\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.471475 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs" (OuterVolumeSpecName: "logs") pod "403c7fd8-7d16-4f89-89f7-f10ec283553c" (UID: "403c7fd8-7d16-4f89-89f7-f10ec283553c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.472212 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/403c7fd8-7d16-4f89-89f7-f10ec283553c-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.479985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8" (OuterVolumeSpecName: "kube-api-access-swdq8") pod "403c7fd8-7d16-4f89-89f7-f10ec283553c" (UID: "403c7fd8-7d16-4f89-89f7-f10ec283553c"). InnerVolumeSpecName "kube-api-access-swdq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.518184 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "403c7fd8-7d16-4f89-89f7-f10ec283553c" (UID: "403c7fd8-7d16-4f89-89f7-f10ec283553c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.521415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data" (OuterVolumeSpecName: "config-data") pod "403c7fd8-7d16-4f89-89f7-f10ec283553c" (UID: "403c7fd8-7d16-4f89-89f7-f10ec283553c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.564357 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "403c7fd8-7d16-4f89-89f7-f10ec283553c" (UID: "403c7fd8-7d16-4f89-89f7-f10ec283553c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.574926 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.574966 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.574980 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/403c7fd8-7d16-4f89-89f7-f10ec283553c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.574992 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swdq8\" (UniqueName: \"kubernetes.io/projected/403c7fd8-7d16-4f89-89f7-f10ec283553c-kube-api-access-swdq8\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.763319 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.770232 4770 generic.go:334] "Generic (PLEG): container finished" podID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerID="e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b" exitCode=0 Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.770376 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.771358 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerDied","Data":"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b"} Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.771489 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"403c7fd8-7d16-4f89-89f7-f10ec283553c","Type":"ContainerDied","Data":"4d6f10d7b0476fb623ef284a1c993e029638abd2f909dd2936b6afc6ea186465"} Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.771581 4770 scope.go:117] "RemoveContainer" containerID="e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.773886 4770 generic.go:334] "Generic (PLEG): container finished" podID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" exitCode=0 Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.774032 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3","Type":"ContainerDied","Data":"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992"} Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.774111 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3","Type":"ContainerDied","Data":"4fa0d07083f52393eef891b1bfc559439e1d32adae87cdbebb79f801cbd7e6d6"} Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.774237 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.859701 4770 scope.go:117] "RemoveContainer" containerID="0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.859809 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.881713 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data\") pod \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.881853 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn8fw\" (UniqueName: \"kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw\") pod \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.882020 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle\") pod \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\" (UID: \"4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3\") " Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.891308 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.891496 4770 scope.go:117] "RemoveContainer" containerID="e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.892094 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b\": container with ID starting with e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b not found: ID does not exist" containerID="e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.892154 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b"} err="failed to get container status \"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b\": rpc error: code = NotFound desc = could not find container \"e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b\": container with ID starting with e7dd3fad815d698ff5700ae03bb52864ecb08e284d744632d4e1438e7ede0f3b not found: ID does not exist" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.892188 4770 scope.go:117] "RemoveContainer" containerID="0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.894317 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7\": container with ID starting with 0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7 not found: ID does not exist" containerID="0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.894388 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7"} err="failed to get container status \"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7\": rpc error: code = NotFound desc = could not find container \"0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7\": container with ID starting with 0849076c1d603ebe6decb081e54e5a944770cc97a1b805b1060174ef6fc894c7 not found: ID does not exist" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.894424 4770 scope.go:117] "RemoveContainer" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.895171 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw" (OuterVolumeSpecName: "kube-api-access-zn8fw") pod "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" (UID: "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3"). InnerVolumeSpecName "kube-api-access-zn8fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.907259 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.907870 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.907911 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.907930 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.907941 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.907965 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.907974 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.908229 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-metadata" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.908256 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" containerName="nova-metadata-log" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.908280 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" containerName="nova-scheduler-scheduler" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.909764 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.912934 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.913073 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.919679 4770 scope.go:117] "RemoveContainer" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" Dec 09 11:57:54 crc kubenswrapper[4770]: E1209 11:57:54.920366 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992\": container with ID starting with dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992 not found: ID does not exist" containerID="dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.920427 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992"} err="failed to get container status \"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992\": rpc error: code = NotFound desc = could not find container \"dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992\": container with ID starting with dd6fef2feb72c8e10584f145be45a678e22ab3c72acda6f1c64766d949329992 not found: ID does not exist" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.924469 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.965410 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data" (OuterVolumeSpecName: "config-data") pod "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" (UID: "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.971088 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" (UID: "4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.993135 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.993187 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn8fw\" (UniqueName: \"kubernetes.io/projected/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-kube-api-access-zn8fw\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:54 crc kubenswrapper[4770]: I1209 11:57:54.993203 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.094623 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.094986 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8486\" (UniqueName: \"kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.095037 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.095153 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.095340 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.114384 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.126090 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.143259 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.145117 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.151580 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.163480 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.197270 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.197353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.197433 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.197479 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8486\" (UniqueName: \"kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.197524 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.198064 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.201099 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.201378 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.201534 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.225499 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8486\" (UniqueName: \"kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486\") pod \"nova-metadata-0\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.299580 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hfdh\" (UniqueName: \"kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.299997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.300778 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.352181 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.356711 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3" path="/var/lib/kubelet/pods/4011c8d4-9c15-4d9a-a4e8-0b15ba2a5ec3/volumes" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.357470 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="403c7fd8-7d16-4f89-89f7-f10ec283553c" path="/var/lib/kubelet/pods/403c7fd8-7d16-4f89-89f7-f10ec283553c/volumes" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.403283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hfdh\" (UniqueName: \"kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.403600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.403768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.408419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.409836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.426589 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hfdh\" (UniqueName: \"kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh\") pod \"nova-scheduler-0\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.466920 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:57:55 crc kubenswrapper[4770]: I1209 11:57:55.821450 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.017356 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:57:56 crc kubenswrapper[4770]: W1209 11:57:56.033348 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcead551d_3523_499a_a1b4_ebce9c97e59e.slice/crio-94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3 WatchSource:0}: Error finding container 94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3: Status 404 returned error can't find the container with id 94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3 Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.807882 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cead551d-3523-499a-a1b4-ebce9c97e59e","Type":"ContainerStarted","Data":"2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8"} Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.808298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cead551d-3523-499a-a1b4-ebce9c97e59e","Type":"ContainerStarted","Data":"94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3"} Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.811992 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerStarted","Data":"603f5bb321d03410442e30b0c6f4a276f62b4480a2b95aa1ef1c48a81bca45e8"} Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.812016 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerStarted","Data":"b18e180734f74a73d57257fdf1cd1f9761bbb9036ddc64718a4744fc937a04d5"} Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.812029 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerStarted","Data":"859d1a8633f3d060ab714d46ef835e7532895cb15b70431606765584fe929fbf"} Dec 09 11:57:56 crc kubenswrapper[4770]: I1209 11:57:56.870427 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.8704059499999999 podStartE2EDuration="1.87040595s" podCreationTimestamp="2025-12-09 11:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:56.83735725 +0000 UTC m=+1602.078115769" watchObservedRunningTime="2025-12-09 11:57:56.87040595 +0000 UTC m=+1602.111164469" Dec 09 11:58:00 crc kubenswrapper[4770]: I1209 11:58:00.352753 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:58:00 crc kubenswrapper[4770]: I1209 11:58:00.353384 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 11:58:00 crc kubenswrapper[4770]: I1209 11:58:00.468398 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.154940 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.155006 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.518258 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=7.518238736 podStartE2EDuration="7.518238736s" podCreationTimestamp="2025-12-09 11:57:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:57:56.87318048 +0000 UTC m=+1602.113938999" watchObservedRunningTime="2025-12-09 11:58:01.518238736 +0000 UTC m=+1606.758997255" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.531178 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.536885 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.549461 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.632274 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzk47\" (UniqueName: \"kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.632331 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.632496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.734823 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzk47\" (UniqueName: \"kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.734917 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.735154 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.735592 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.735716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.768326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzk47\" (UniqueName: \"kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47\") pod \"community-operators-h89x8\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:01 crc kubenswrapper[4770]: I1209 11:58:01.864390 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.167160 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.167422 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.473396 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.473457 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.519278 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:02 crc kubenswrapper[4770]: W1209 11:58:02.528517 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4342f43_36bd_4e89_b6a1_a8637443f5e4.slice/crio-9186f3983668580e6d2d85ba5320ce9215809f44d9289cbeefad5b322700ab55 WatchSource:0}: Error finding container 9186f3983668580e6d2d85ba5320ce9215809f44d9289cbeefad5b322700ab55: Status 404 returned error can't find the container with id 9186f3983668580e6d2d85ba5320ce9215809f44d9289cbeefad5b322700ab55 Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.908970 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerID="6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e" exitCode=0 Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.909042 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerDied","Data":"6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e"} Dec 09 11:58:02 crc kubenswrapper[4770]: I1209 11:58:02.909340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerStarted","Data":"9186f3983668580e6d2d85ba5320ce9215809f44d9289cbeefad5b322700ab55"} Dec 09 11:58:03 crc kubenswrapper[4770]: I1209 11:58:03.921326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerStarted","Data":"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd"} Dec 09 11:58:04 crc kubenswrapper[4770]: I1209 11:58:04.933962 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerID="49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd" exitCode=0 Dec 09 11:58:04 crc kubenswrapper[4770]: I1209 11:58:04.934022 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerDied","Data":"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd"} Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.353546 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.353877 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.468835 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.505124 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.946962 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerStarted","Data":"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e"} Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.984537 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h89x8" podStartSLOduration=2.573086953 podStartE2EDuration="4.984517634s" podCreationTimestamp="2025-12-09 11:58:01 +0000 UTC" firstStartedPulling="2025-12-09 11:58:02.912640263 +0000 UTC m=+1608.153398782" lastFinishedPulling="2025-12-09 11:58:05.324070944 +0000 UTC m=+1610.564829463" observedRunningTime="2025-12-09 11:58:05.979725933 +0000 UTC m=+1611.220484452" watchObservedRunningTime="2025-12-09 11:58:05.984517634 +0000 UTC m=+1611.225276153" Dec 09 11:58:05 crc kubenswrapper[4770]: I1209 11:58:05.984638 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 11:58:06 crc kubenswrapper[4770]: I1209 11:58:06.365183 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:06 crc kubenswrapper[4770]: I1209 11:58:06.365214 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:08 crc kubenswrapper[4770]: I1209 11:58:08.457234 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="83f8d015-ef90-4b96-ad91-8c6ffa850a25" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.189:3000/\": dial tcp 10.217.0.189:3000: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:09 crc kubenswrapper[4770]: I1209 11:58:09.901198 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.165438 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.165878 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.166361 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.166427 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.175431 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.177572 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.865360 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.865700 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:11 crc kubenswrapper[4770]: I1209 11:58:11.925613 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:12 crc kubenswrapper[4770]: I1209 11:58:12.349166 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:12 crc kubenswrapper[4770]: I1209 11:58:12.421462 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:14 crc kubenswrapper[4770]: I1209 11:58:14.299627 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h89x8" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="registry-server" containerID="cri-o://3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e" gracePeriod=2 Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.771637 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.857214 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities\") pod \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.967300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities" (OuterVolumeSpecName: "utilities") pod "e4342f43-36bd-4e89-b6a1-a8637443f5e4" (UID: "e4342f43-36bd-4e89-b6a1-a8637443f5e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.968341 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content\") pod \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.968890 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzk47\" (UniqueName: \"kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47\") pod \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\" (UID: \"e4342f43-36bd-4e89-b6a1-a8637443f5e4\") " Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.969341 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:14.983586 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47" (OuterVolumeSpecName: "kube-api-access-hzk47") pod "e4342f43-36bd-4e89-b6a1-a8637443f5e4" (UID: "e4342f43-36bd-4e89-b6a1-a8637443f5e4"). InnerVolumeSpecName "kube-api-access-hzk47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.058447 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4342f43-36bd-4e89-b6a1-a8637443f5e4" (UID: "e4342f43-36bd-4e89-b6a1-a8637443f5e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.071446 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzk47\" (UniqueName: \"kubernetes.io/projected/e4342f43-36bd-4e89-b6a1-a8637443f5e4-kube-api-access-hzk47\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.071484 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4342f43-36bd-4e89-b6a1-a8637443f5e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.318646 4770 generic.go:334] "Generic (PLEG): container finished" podID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerID="3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e" exitCode=0 Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.318710 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerDied","Data":"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e"} Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.318739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89x8" event={"ID":"e4342f43-36bd-4e89-b6a1-a8637443f5e4","Type":"ContainerDied","Data":"9186f3983668580e6d2d85ba5320ce9215809f44d9289cbeefad5b322700ab55"} Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.318742 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89x8" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.318774 4770 scope.go:117] "RemoveContainer" containerID="3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.360788 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.361559 4770 scope.go:117] "RemoveContainer" containerID="49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.368574 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.369053 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.373575 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h89x8"] Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.380000 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.380265 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.389671 4770 scope.go:117] "RemoveContainer" containerID="6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.451599 4770 scope.go:117] "RemoveContainer" containerID="3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e" Dec 09 11:58:15 crc kubenswrapper[4770]: E1209 11:58:15.452565 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e\": container with ID starting with 3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e not found: ID does not exist" containerID="3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.452615 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e"} err="failed to get container status \"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e\": rpc error: code = NotFound desc = could not find container \"3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e\": container with ID starting with 3cc1e96ca3448bc9c31ee5f16d75d0dca102925360b783fa2bd9e08c6f8aac8e not found: ID does not exist" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.452649 4770 scope.go:117] "RemoveContainer" containerID="49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd" Dec 09 11:58:15 crc kubenswrapper[4770]: E1209 11:58:15.452915 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd\": container with ID starting with 49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd not found: ID does not exist" containerID="49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.452948 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd"} err="failed to get container status \"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd\": rpc error: code = NotFound desc = could not find container \"49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd\": container with ID starting with 49e8e20036b340b6746cc7a2186452a11816af65b7dcc83d107ac7b10a57b9fd not found: ID does not exist" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.452972 4770 scope.go:117] "RemoveContainer" containerID="6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e" Dec 09 11:58:15 crc kubenswrapper[4770]: E1209 11:58:15.453769 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e\": container with ID starting with 6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e not found: ID does not exist" containerID="6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e" Dec 09 11:58:15 crc kubenswrapper[4770]: I1209 11:58:15.453792 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e"} err="failed to get container status \"6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e\": rpc error: code = NotFound desc = could not find container \"6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e\": container with ID starting with 6191d2d450c6167778215ab50420fe10e6fe706fe37e946b5d56f46d119de60e not found: ID does not exist" Dec 09 11:58:17 crc kubenswrapper[4770]: I1209 11:58:17.350554 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" path="/var/lib/kubelet/pods/e4342f43-36bd-4e89-b6a1-a8637443f5e4/volumes" Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.474281 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.474800 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.474853 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.475653 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.475707 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" gracePeriod=600 Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.831968 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" exitCode=0 Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.832050 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118"} Dec 09 11:58:32 crc kubenswrapper[4770]: I1209 11:58:32.832648 4770 scope.go:117] "RemoveContainer" containerID="2502940d245941b395a0724de963b4ad18304d20fe1e64150d9c63928b83a95f" Dec 09 11:58:33 crc kubenswrapper[4770]: E1209 11:58:33.201773 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:58:33 crc kubenswrapper[4770]: I1209 11:58:33.611528 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:58:33 crc kubenswrapper[4770]: E1209 11:58:33.847522 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:33 crc kubenswrapper[4770]: E1209 11:58:33.847588 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data podName:b596f863-b8d8-4fb4-93e3-45bdce78f1b1 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:34.347571522 +0000 UTC m=+1639.588330041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data") pod "rabbitmq-cell1-server-0" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1") : configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:33 crc kubenswrapper[4770]: I1209 11:58:33.931327 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:58:33 crc kubenswrapper[4770]: I1209 11:58:33.934841 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 09 11:58:33 crc kubenswrapper[4770]: I1209 11:58:33.935355 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" containerName="openstackclient" containerID="cri-o://faf973a0c6d5bcf16a241dc80a13699d03f0a874da0549eb1f5e4001556d13ae" gracePeriod=2 Dec 09 11:58:33 crc kubenswrapper[4770]: E1209 11:58:33.934966 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:58:33 crc kubenswrapper[4770]: I1209 11:58:33.982223 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.106912 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.107779 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="openstack-network-exporter" containerID="cri-o://7b634173148f824de10487babb4e2f965f1000bfa7ef0c55f90a98000dbccbff" gracePeriod=300 Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.142434 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.271680 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.271741 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data podName:03fef9de-3f78-48b3-9079-9dc87184f803 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:34.771728071 +0000 UTC m=+1640.012486590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data") pod "rabbitmq-server-0" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803") : configmap "rabbitmq-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.312585 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="ovsdbserver-sb" containerID="cri-o://f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" gracePeriod=300 Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.355180 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement6c0f-account-delete-j5h7v"] Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.355767 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="registry-server" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.355785 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="registry-server" Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.355801 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="extract-utilities" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.355811 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="extract-utilities" Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.355832 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="extract-content" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.355838 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="extract-content" Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.355855 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" containerName="openstackclient" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.355861 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" containerName="openstackclient" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.356080 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4342f43-36bd-4e89-b6a1-a8637443f5e4" containerName="registry-server" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.356101 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" containerName="openstackclient" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.356887 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.370396 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.370683 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="openstack-network-exporter" containerID="cri-o://5525f70a6a808175e62c919c6f6e56ec54293e10c519fefa5b6d1b5a45deea6e" gracePeriod=300 Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.372763 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.372824 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data podName:b596f863-b8d8-4fb4-93e3-45bdce78f1b1 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:35.372810208 +0000 UTC m=+1640.613568727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data") pod "rabbitmq-cell1-server-0" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1") : configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.413146 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement6c0f-account-delete-j5h7v"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.475190 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql2j4\" (UniqueName: \"kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.475352 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.519421 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.529629 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.568409 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.582858 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql2j4\" (UniqueName: \"kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.583002 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.583886 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.632339 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="ovsdbserver-nb" containerID="cri-o://3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" gracePeriod=300 Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.644159 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanaaaf-account-delete-x5lgz"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.645681 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.676592 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql2j4\" (UniqueName: \"kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4\") pod \"placement6c0f-account-delete-j5h7v\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.689976 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcdd5\" (UniqueName: \"kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.690089 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.697171 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.703668 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanaaaf-account-delete-x5lgz"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.775974 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zjjm9"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.793140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.793294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcdd5\" (UniqueName: \"kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.793352 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.793438 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq2xz\" (UniqueName: \"kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.793617 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: E1209 11:58:34.793669 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data podName:03fef9de-3f78-48b3-9079-9dc87184f803 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:35.793652504 +0000 UTC m=+1641.034411023 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data") pod "rabbitmq-server-0" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803") : configmap "rabbitmq-config-data" not found Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.794744 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.814303 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zjjm9"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.852444 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcdd5\" (UniqueName: \"kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5\") pod \"cinder4e4b-account-delete-5tttq\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.873677 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.874029 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" containerID="cri-o://867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" gracePeriod=30 Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.874538 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="openstack-network-exporter" containerID="cri-o://60b3561b63896f5bc9807fe5e6de66a0dadd48252dfe9a1b9a2f39c48715e956" gracePeriod=30 Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.895794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq2xz\" (UniqueName: \"kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.895856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.896758 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.913671 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.915411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.934528 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:34 crc kubenswrapper[4770]: I1209 11:58:34.983018 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq2xz\" (UniqueName: \"kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz\") pod \"barbicanaaaf-account-delete-x5lgz\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.004677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh864\" (UniqueName: \"kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.004941 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.022799 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.035513 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.052335 4770 generic.go:334] "Generic (PLEG): container finished" podID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerID="5525f70a6a808175e62c919c6f6e56ec54293e10c519fefa5b6d1b5a45deea6e" exitCode=2 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.052419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerDied","Data":"5525f70a6a808175e62c919c6f6e56ec54293e10c519fefa5b6d1b5a45deea6e"} Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.106678 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh864\" (UniqueName: \"kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.107047 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.107977 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.131538 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.147061 4770 generic.go:334] "Generic (PLEG): container finished" podID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerID="7b634173148f824de10487babb4e2f965f1000bfa7ef0c55f90a98000dbccbff" exitCode=2 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.147118 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerDied","Data":"7b634173148f824de10487babb4e2f965f1000bfa7ef0c55f90a98000dbccbff"} Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.178118 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh864\" (UniqueName: \"kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864\") pod \"neutron2ce4-account-delete-z5lmd\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.216668 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.216967 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-jfg8z" podUID="cabab723-8add-4dd0-b0cc-551f30a039d3" containerName="openstack-network-exporter" containerID="cri-o://33128ce32a57707b0adfe4148bebb6dd61f023516c99ba4c5d553f50864934f7" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.268282 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.309531 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4m5vj"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.335508 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4m5vj"] Dec 09 11:58:35 crc kubenswrapper[4770]: E1209 11:58:35.421629 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:35 crc kubenswrapper[4770]: E1209 11:58:35.422040 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data podName:b596f863-b8d8-4fb4-93e3-45bdce78f1b1 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:37.42201831 +0000 UTC m=+1642.662776829 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data") pod "rabbitmq-cell1-server-0" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1") : configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.444716 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea745621-9b83-4af8-bb18-7fea76a4167d" path="/var/lib/kubelet/pods/ea745621-9b83-4af8-bb18-7fea76a4167d/volumes" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.451172 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2433461-ff2f-4364-adca-d649cd8d68a9" path="/var/lib/kubelet/pods/f2433461-ff2f-4364-adca-d649cd8d68a9/volumes" Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.452115 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.452142 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-v7q9j"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.497003 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-v7q9j"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.804382 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-p2vmw"] Dec 09 11:58:35 crc kubenswrapper[4770]: E1209 11:58:35.816337 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 09 11:58:35 crc kubenswrapper[4770]: E1209 11:58:35.816510 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data podName:03fef9de-3f78-48b3-9079-9dc87184f803 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:37.816489846 +0000 UTC m=+1643.057248365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data") pod "rabbitmq-server-0" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803") : configmap "rabbitmq-config-data" not found Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.831297 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-p2vmw"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.853987 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855069 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-server" containerID="cri-o://d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855216 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-updater" containerID="cri-o://4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855219 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-server" containerID="cri-o://4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855287 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-auditor" containerID="cri-o://4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855358 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-replicator" containerID="cri-o://9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855402 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-server" containerID="cri-o://14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855443 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-reaper" containerID="cri-o://d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855479 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-auditor" containerID="cri-o://ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855522 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-replicator" containerID="cri-o://87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.855873 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-expirer" containerID="cri-o://80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.856318 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-updater" containerID="cri-o://5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.856410 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-auditor" containerID="cri-o://3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.856459 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-replicator" containerID="cri-o://7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.856540 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="rsync" containerID="cri-o://13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.856013 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="swift-recon-cron" containerID="cri-o://e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6" gracePeriod=30 Dec 09 11:58:35 crc kubenswrapper[4770]: I1209 11:58:35.908281 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-l9l6m"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.009210 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-pxdc5"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.059921 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-l9l6m"] Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.086517 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b is running failed: container process not found" containerID="3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.089567 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b is running failed: container process not found" containerID="3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.092251 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b is running failed: container process not found" containerID="3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.092326 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="ovsdbserver-nb" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.112749 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.114395 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.140500 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-pxdc5"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.188717 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.189109 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="dnsmasq-dns" containerID="cri-o://43ebc26e2c2fa580b48c4844f5bfe5b3ba67d22ccd82e4aaaa68755ef5a9c123" gracePeriod=10 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.212823 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8983440f-2acf-40e0-a2fb-757413c0d0bb/ovsdbserver-sb/0.log" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.212883 4770 generic.go:334] "Generic (PLEG): container finished" podID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerID="f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" exitCode=143 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.212982 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerDied","Data":"f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.214175 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.233546 4770 generic.go:334] "Generic (PLEG): container finished" podID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerID="60b3561b63896f5bc9807fe5e6de66a0dadd48252dfe9a1b9a2f39c48715e956" exitCode=2 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.233678 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerDied","Data":"60b3561b63896f5bc9807fe5e6de66a0dadd48252dfe9a1b9a2f39c48715e956"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.238613 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.245626 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jfg8z_cabab723-8add-4dd0-b0cc-551f30a039d3/openstack-network-exporter/0.log" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.245686 4770 generic.go:334] "Generic (PLEG): container finished" podID="cabab723-8add-4dd0-b0cc-551f30a039d3" containerID="33128ce32a57707b0adfe4148bebb6dd61f023516c99ba4c5d553f50864934f7" exitCode=2 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.245796 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jfg8z" event={"ID":"cabab723-8add-4dd0-b0cc-551f30a039d3","Type":"ContainerDied","Data":"33128ce32a57707b0adfe4148bebb6dd61f023516c99ba4c5d553f50864934f7"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.250598 4770 generic.go:334] "Generic (PLEG): container finished" podID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" containerID="faf973a0c6d5bcf16a241dc80a13699d03f0a874da0549eb1f5e4001556d13ae" exitCode=137 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.253032 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzpmn\" (UniqueName: \"kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.253142 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.258395 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jb49v"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.261388 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_518c58bd-4a2f-439f-b5f5-0eda21434884/ovsdbserver-nb/0.log" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.272003 4770 generic.go:334] "Generic (PLEG): container finished" podID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerID="3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" exitCode=143 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.272130 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerDied","Data":"3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.276981 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.283489 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-574ccd49f8-67m27" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-log" containerID="cri-o://4ba87b24c19c1120bbcabbace77c866035e530eaf47c87572eba2d7e47526454" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.284007 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-574ccd49f8-67m27" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-api" containerID="cri-o://b505b07675385a501eefd16b47033f1d0cdb0b120345d67e47e86e33ef0f2548" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.306998 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-qzmrd"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.322711 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" containerID="cri-o://f2d4dc50bf2b36a841745128ed7fc9d07db8df1ab56a2af6bc7a83543d7f633f" gracePeriod=604800 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.327291 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-qzmrd"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.349359 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jb49v"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.360281 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzpmn\" (UniqueName: \"kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.362768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.363978 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371223 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371269 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371281 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371289 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371297 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371307 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371315 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371326 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371337 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371346 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d" exitCode=0 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371394 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371463 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371476 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371488 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371500 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371512 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371522 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371532 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371546 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d"} Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.371855 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="cinder-scheduler" containerID="cri-o://d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.372024 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="probe" containerID="cri-o://7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.397700 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.406461 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.406779 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api-log" containerID="cri-o://7bc283a32d2b16f74f1cf489c744ab75e9f7ab17baf7e2b5ac315e76c1b74b42" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.406935 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api" containerID="cri-o://28cd7242a9c9640679276453e171ec22b26b2720ffe571f2f85b629ee46a5e06" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.423569 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.423970 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dc5b689b7-qgl92" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-api" containerID="cri-o://fd8527bc2e276f42630b31d0e4c2d54c775ef3cdb370d0ca709fc11bfc05e843" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.424695 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dc5b689b7-qgl92" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-httpd" containerID="cri-o://85925850bf61e5c8f64a81b9f9e5037d619b24f1d0912e5127b4a2874bb652ee" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.442359 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell17adb-account-delete-rtcfc"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.444379 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.491373 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" containerID="cri-o://1980ac00c990360135915c858d0f125e136b46aa153c5a4d5cf4019dd84f4355" gracePeriod=604800 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.491544 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.491792 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-log" containerID="cri-o://7569e5bc00921bcf0e4cb0006c71b05ba900b55b24a670f35b3c7271e62f80ca" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.491882 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-httpd" containerID="cri-o://eebed45a512ca372586490859e850a999b9d581d64647896a65dfb070e1af5fb" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.528387 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzpmn\" (UniqueName: \"kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn\") pod \"glance078e-account-delete-lwfnm\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.538650 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell17adb-account-delete-rtcfc"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.577505 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.577819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.616237 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.616606 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener-log" containerID="cri-o://e77c216dc47b6c7c9a614efb4c3cf8153cbd683bc99a04574e90b93b09aa208a" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.617294 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener" containerID="cri-o://bf8ad42774200d477849ea268cd29b0c65b5c006ef52af1b6710676014a7e842" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.673134 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.675965 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.676268 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66bfc44b69-qfbmc" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker-log" containerID="cri-o://c5ec8fa6eed14138a68a6b229d128065bd09d1ea7c60899215996437c5d9b880" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.676393 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66bfc44b69-qfbmc" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker" containerID="cri-o://3390dfb8cbb1f0c6bd1a75c816ce715ca36a3164d095376f2127e25ddb826b67" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.676995 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.679596 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.679967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.680235 4770 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.680306 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:37.18028818 +0000 UTC m=+1642.421046699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : configmap "openstack-cell1-scripts" not found Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.680532 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.680650 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.685009 4770 projected.go:194] Error preparing data for projected volume kube-api-access-5dpw7 for pod openstack/novacell17adb-account-delete-rtcfc: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:36 crc kubenswrapper[4770]: E1209 11:58:36.685266 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7 podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:37.185247766 +0000 UTC m=+1642.426006285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5dpw7" (UniqueName: "kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.689407 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.691217 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.714027 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.726877 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi8c24-account-delete-ccctw"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.728671 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.774383 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.774750 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-log" containerID="cri-o://243be21689d2a29c2966994727f43b6692216066e84a9a1e11711c1c2553ffb8" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.775289 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-api" containerID="cri-o://9a473cbcc51156e661f62a9cb155beb880dd96bd5e5e0ee00faf14c8c0beeb93" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.783103 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.783137 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czxht\" (UniqueName: \"kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.805043 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.822637 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi8c24-account-delete-ccctw"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.831958 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.832249 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-log" containerID="cri-o://8ce8e611661071a5e49fc120f1a9bef7f615630c825216f4eed7d8cb555e5d6c" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.832467 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-httpd" containerID="cri-o://0193e4873f3244cab191ab4be299640f973319dfcf035f612a78cdfcd54c17a7" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.853029 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.853329 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" containerID="cri-o://b18e180734f74a73d57257fdf1cd1f9761bbb9036ddc64718a4744fc937a04d5" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.853890 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" containerID="cri-o://603f5bb321d03410442e30b0c6f4a276f62b4480a2b95aa1ef1c48a81bca45e8" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.873043 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" containerID="cri-o://5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" gracePeriod=29 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.884864 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29chb\" (UniqueName: \"kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.885014 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.885205 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.885278 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czxht\" (UniqueName: \"kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.886296 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.918464 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.920118 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czxht\" (UniqueName: \"kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht\") pod \"novacell017c1-account-delete-xwp6r\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.974428 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.974653 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" containerID="cri-o://0ed885e5102a41a0b21a56c6ea93cb492355c4740fed766c91160632cf4c5c43" gracePeriod=30 Dec 09 11:58:36 crc kubenswrapper[4770]: I1209 11:58:36.977157 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" containerID="cri-o://9b65050fb96dad2b68de130b2e74efdc4a249da8425436217ee12e2727d54d98" gracePeriod=30 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:36.988788 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29chb\" (UniqueName: \"kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:36.988848 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:36.999765 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.009487 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.009770 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130" gracePeriod=30 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.073758 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xhvz6"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.077860 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29chb\" (UniqueName: \"kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb\") pod \"novaapi8c24-account-delete-ccctw\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.084799 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.098475 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell17adb-account-delete-rtcfc"] Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.099538 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-5dpw7 operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell17adb-account-delete-rtcfc" podUID="0fccfd7f-e369-460c-86b6-16add156ef57" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.113484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.124483 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jfg8z_cabab723-8add-4dd0-b0cc-551f30a039d3/openstack-network-exporter/0.log" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.124546 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.125542 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xhvz6"] Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.155568 4770 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 09 11:58:37 crc kubenswrapper[4770]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 09 11:58:37 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNBridge=br-int Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Dec 09 11:58:37 crc kubenswrapper[4770]: ++ PhysicalNetworks= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNHostName= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 09 11:58:37 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 09 11:58:37 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Dec 09 11:58:37 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 09 11:58:37 crc kubenswrapper[4770]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-gct95" message=< Dec 09 11:58:37 crc kubenswrapper[4770]: Exiting ovsdb-server (5) [ OK ] Dec 09 11:58:37 crc kubenswrapper[4770]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 09 11:58:37 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNBridge=br-int Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Dec 09 11:58:37 crc kubenswrapper[4770]: ++ PhysicalNetworks= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNHostName= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 09 11:58:37 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 09 11:58:37 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Dec 09 11:58:37 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 09 11:58:37 crc kubenswrapper[4770]: > Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.155626 4770 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 09 11:58:37 crc kubenswrapper[4770]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 09 11:58:37 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNBridge=br-int Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Dec 09 11:58:37 crc kubenswrapper[4770]: ++ PhysicalNetworks= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ OVNHostName= Dec 09 11:58:37 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 09 11:58:37 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 09 11:58:37 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 09 11:58:37 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + sleep 0.5 Dec 09 11:58:37 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 09 11:58:37 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Dec 09 11:58:37 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 09 11:58:37 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 09 11:58:37 crc kubenswrapper[4770]: > pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" containerID="cri-o://37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.155666 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" containerID="cri-o://37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" gracePeriod=28 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.215435 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.215770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqrrh\" (UniqueName: \"kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.215911 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.216017 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.216114 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.216224 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir\") pod \"cabab723-8add-4dd0-b0cc-551f30a039d3\" (UID: \"cabab723-8add-4dd0-b0cc-551f30a039d3\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.217076 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.217266 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.225246 4770 projected.go:194] Error preparing data for projected volume kube-api-access-5dpw7 for pod openstack/novacell17adb-account-delete-rtcfc: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.225351 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7 podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:38.225324659 +0000 UTC m=+1643.466083178 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-5dpw7" (UniqueName: "kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.226044 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.230212 4770 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.230307 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:38.230275854 +0000 UTC m=+1643.471034383 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : configmap "openstack-cell1-scripts" not found Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.230466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.245934 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh" (OuterVolumeSpecName: "kube-api-access-fqrrh") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "kube-api-access-fqrrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.246356 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7adb-account-create-update-nkbct"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.284263 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config" (OuterVolumeSpecName: "config") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.320935 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqrrh\" (UniqueName: \"kubernetes.io/projected/cabab723-8add-4dd0-b0cc-551f30a039d3-kube-api-access-fqrrh\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.320981 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.320996 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cabab723-8add-4dd0-b0cc-551f30a039d3-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.321007 4770 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cabab723-8add-4dd0-b0cc-551f30a039d3-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.375217 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.391419 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67 is running failed: container process not found" containerID="f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.394345 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67 is running failed: container process not found" containerID="f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.397275 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67 is running failed: container process not found" containerID="f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.397343 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="ovsdbserver-sb" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.397559 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c" path="/var/lib/kubelet/pods/07e6b8db-e2c8-49a2-b8b8-49bc8bbe203c/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.407680 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a" path="/var/lib/kubelet/pods/0b6fbee0-ba33-42f7-8fcb-46ed38bfcb8a/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.422638 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c091c39-7a67-414c-9cbd-d323073c87be" path="/var/lib/kubelet/pods/0c091c39-7a67-414c-9cbd-d323073c87be/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.428216 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b963017-2a12-4106-b6da-49b6725bfd9d" path="/var/lib/kubelet/pods/3b963017-2a12-4106-b6da-49b6725bfd9d/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.428952 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508d8422-6ead-4541-b5e3-b16337546e7d" path="/var/lib/kubelet/pods/508d8422-6ead-4541-b5e3-b16337546e7d/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.429568 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fdfdc9d-aa13-4fdb-8e1f-c54801e47867" path="/var/lib/kubelet/pods/9fdfdc9d-aa13-4fdb-8e1f-c54801e47867/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.440247 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8ce6a43-137d-49d2-8f22-d72d01e55ae5" path="/var/lib/kubelet/pods/a8ce6a43-137d-49d2-8f22-d72d01e55ae5/volumes" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.441868 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.442078 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.442136 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data podName:b596f863-b8d8-4fb4-93e3-45bdce78f1b1 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:41.442121074 +0000 UTC m=+1646.682879593 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data") pod "rabbitmq-cell1-server-0" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1") : configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.488461 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="galera" containerID="cri-o://2946bef94b26e9eb5227b621b0591908bb8f5db0d79016f7e9097a651629b7fc" gracePeriod=30 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.506551 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7adb-account-create-update-nkbct"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.506598 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement6c0f-account-delete-j5h7v"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.530584 4770 generic.go:334] "Generic (PLEG): container finished" podID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerID="0ed885e5102a41a0b21a56c6ea93cb492355c4740fed766c91160632cf4c5c43" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.530736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerDied","Data":"0ed885e5102a41a0b21a56c6ea93cb492355c4740fed766c91160632cf4c5c43"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.532883 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.543131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanaaaf-account-delete-x5lgz" event={"ID":"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb","Type":"ContainerStarted","Data":"9e890bcfd1a493e1a8d379b97b5e7d8b67b9f9c84627c7cfdefc69f24605ef06"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.543510 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8983440f-2acf-40e0-a2fb-757413c0d0bb/ovsdbserver-sb/0.log" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.543579 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.650800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.650982 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651038 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651133 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651171 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqfnq\" (UniqueName: \"kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651190 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.651378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts\") pod \"8983440f-2acf-40e0-a2fb-757413c0d0bb\" (UID: \"8983440f-2acf-40e0-a2fb-757413c0d0bb\") " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.653122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts" (OuterVolumeSpecName: "scripts") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.653722 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config" (OuterVolumeSpecName: "config") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.655857 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.672523 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.193:6080/vnc_lite.html\": dial tcp 10.217.0.193:6080: connect: connection refused" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.677948 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.686031 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq" (OuterVolumeSpecName: "kube-api-access-hqfnq") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "kube-api-access-hqfnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.686776 4770 generic.go:334] "Generic (PLEG): container finished" podID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerID="243be21689d2a29c2966994727f43b6692216066e84a9a1e11711c1c2553ffb8" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.687781 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerDied","Data":"243be21689d2a29c2966994727f43b6692216066e84a9a1e11711c1c2553ffb8"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.707388 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.723151 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanaaaf-account-delete-x5lgz"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.732226 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jfg8z_cabab723-8add-4dd0-b0cc-551f30a039d3/openstack-network-exporter/0.log" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.732357 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jfg8z" event={"ID":"cabab723-8add-4dd0-b0cc-551f30a039d3","Type":"ContainerDied","Data":"a5082f92fc1418c0900e41864024dfdfaaf0407045931ba63005855d9fe50f5f"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.732400 4770 scope.go:117] "RemoveContainer" containerID="33128ce32a57707b0adfe4148bebb6dd61f023516c99ba4c5d553f50864934f7" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.732462 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jfg8z" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.768502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder4e4b-account-delete-5tttq" event={"ID":"24be0696-be7e-46b9-aa90-d83d8abbd793","Type":"ContainerStarted","Data":"dc4d3f4170827dac7881677505a31fbd90cfd992ad89a750ab423222c74bfc3f"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.768846 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.769533 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.769561 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.769571 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqfnq\" (UniqueName: \"kubernetes.io/projected/8983440f-2acf-40e0-a2fb-757413c0d0bb-kube-api-access-hqfnq\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.769596 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.769605 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8983440f-2acf-40e0-a2fb-757413c0d0bb-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.815457 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8983440f-2acf-40e0-a2fb-757413c0d0bb/ovsdbserver-sb/0.log" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.816553 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8983440f-2acf-40e0-a2fb-757413c0d0bb","Type":"ContainerDied","Data":"fc5566166462f989c3e85ae3adf50733b43e9d8583eb013a176ba72ae6dfbff5"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.816713 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.896154 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 09 11:58:37 crc kubenswrapper[4770]: E1209 11:58:37.896229 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data podName:03fef9de-3f78-48b3-9079-9dc87184f803 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:41.896209614 +0000 UTC m=+1647.136968133 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data") pod "rabbitmq-server-0" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803") : configmap "rabbitmq-config-data" not found Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.897520 4770 generic.go:334] "Generic (PLEG): container finished" podID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerID="7bc283a32d2b16f74f1cf489c744ab75e9f7ab17baf7e2b5ac315e76c1b74b42" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.897597 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerDied","Data":"7bc283a32d2b16f74f1cf489c744ab75e9f7ab17baf7e2b5ac315e76c1b74b42"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.942970 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.943595 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerName="nova-scheduler-scheduler" containerID="cri-o://2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" gracePeriod=30 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.950585 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder4e4b-account-delete-5tttq" podStartSLOduration=3.950558984 podStartE2EDuration="3.950558984s" podCreationTimestamp="2025-12-09 11:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:58:37.846456421 +0000 UTC m=+1643.087214930" watchObservedRunningTime="2025-12-09 11:58:37.950558984 +0000 UTC m=+1643.191317503" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.951462 4770 generic.go:334] "Generic (PLEG): container finished" podID="9f7b544a-d135-44fc-9fda-3938369f661f" containerID="e77c216dc47b6c7c9a614efb4c3cf8153cbd683bc99a04574e90b93b09aa208a" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.951594 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerDied","Data":"e77c216dc47b6c7c9a614efb4c3cf8153cbd683bc99a04574e90b93b09aa208a"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.963132 4770 scope.go:117] "RemoveContainer" containerID="7b634173148f824de10487babb4e2f965f1000bfa7ef0c55f90a98000dbccbff" Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.965265 4770 generic.go:334] "Generic (PLEG): container finished" podID="568ef04e-032c-40f4-97b9-b21483281a4f" containerID="c5ec8fa6eed14138a68a6b229d128065bd09d1ea7c60899215996437c5d9b880" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.965344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerDied","Data":"c5ec8fa6eed14138a68a6b229d128065bd09d1ea7c60899215996437c5d9b880"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.972759 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8twmk"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.980427 4770 generic.go:334] "Generic (PLEG): container finished" podID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerID="b18e180734f74a73d57257fdf1cd1f9761bbb9036ddc64718a4744fc937a04d5" exitCode=143 Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.980559 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerDied","Data":"b18e180734f74a73d57257fdf1cd1f9761bbb9036ddc64718a4744fc937a04d5"} Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.983987 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8twmk"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.990151 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:58:37 crc kubenswrapper[4770]: I1209 11:58:37.990413 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" containerID="cri-o://96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" gracePeriod=30 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.005055 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.005323 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" containerID="cri-o://5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" gracePeriod=30 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.016565 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qdtv7"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.033938 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qdtv7"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.053685 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerID="8ce8e611661071a5e49fc120f1a9bef7f615630c825216f4eed7d8cb555e5d6c" exitCode=143 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.053812 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerDied","Data":"8ce8e611661071a5e49fc120f1a9bef7f615630c825216f4eed7d8cb555e5d6c"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.108279 4770 generic.go:334] "Generic (PLEG): container finished" podID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerID="85925850bf61e5c8f64a81b9f9e5037d619b24f1d0912e5127b4a2874bb652ee" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.108432 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerDied","Data":"85925850bf61e5c8f64a81b9f9e5037d619b24f1d0912e5127b4a2874bb652ee"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.145836 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.173219 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179286 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179333 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179345 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179354 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179467 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.179477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.196142 4770 generic.go:334] "Generic (PLEG): container finished" podID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerID="43ebc26e2c2fa580b48c4844f5bfe5b3ba67d22ccd82e4aaaa68755ef5a9c123" exitCode=0 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.196306 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" event={"ID":"de2cc3bf-23a6-4b3f-a1fa-b845aad71787","Type":"ContainerDied","Data":"43ebc26e2c2fa580b48c4844f5bfe5b3ba67d22ccd82e4aaaa68755ef5a9c123"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.213927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement6c0f-account-delete-j5h7v" event={"ID":"43a8a640-d610-4726-b2c6-84e99621d820","Type":"ContainerStarted","Data":"b72abc12a6683d231d24f05fd066d588593a15880270b01d02054a0969cc24af"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.213995 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement6c0f-account-delete-j5h7v" event={"ID":"43a8a640-d610-4726-b2c6-84e99621d820","Type":"ContainerStarted","Data":"3442fb69865baa3d225d215aefb85c5bdf397a7a1f6e869fd463140e7857c2fd"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.228518 4770 generic.go:334] "Generic (PLEG): container finished" podID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerID="4ba87b24c19c1120bbcabbace77c866035e530eaf47c87572eba2d7e47526454" exitCode=143 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.228610 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerDied","Data":"4ba87b24c19c1120bbcabbace77c866035e530eaf47c87572eba2d7e47526454"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.244970 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:38 crc kubenswrapper[4770]: E1209 11:58:38.245146 4770 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 09 11:58:38 crc kubenswrapper[4770]: E1209 11:58:38.245380 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:40.24535893 +0000 UTC m=+1645.486117449 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : configmap "openstack-cell1-scripts" not found Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.245638 4770 generic.go:334] "Generic (PLEG): container finished" podID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerID="7569e5bc00921bcf0e4cb0006c71b05ba900b55b24a670f35b3c7271e62f80ca" exitCode=143 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.245782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerDied","Data":"7569e5bc00921bcf0e4cb0006c71b05ba900b55b24a670f35b3c7271e62f80ca"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.245795 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.245820 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.257694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.258524 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2ce4-account-delete-z5lmd" event={"ID":"234ed53e-6da0-4fb5-8990-c4af4f500af8","Type":"ContainerStarted","Data":"3eec329e3fd54ef56513a36e5cdc372b32f3a824e2bb2c5cffd5857e1d917987"} Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.298356 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "cabab723-8add-4dd0-b0cc-551f30a039d3" (UID: "cabab723-8add-4dd0-b0cc-551f30a039d3"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.347243 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.351004 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabab723-8add-4dd0-b0cc-551f30a039d3-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.351383 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:38 crc kubenswrapper[4770]: E1209 11:58:38.355583 4770 projected.go:194] Error preparing data for projected volume kube-api-access-5dpw7 for pod openstack/novacell17adb-account-delete-rtcfc: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:38 crc kubenswrapper[4770]: E1209 11:58:38.355664 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7 podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:40.35564177 +0000 UTC m=+1645.596400289 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-5dpw7" (UniqueName: "kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.407720 4770 scope.go:117] "RemoveContainer" containerID="f09b979e1f82f9a0eb25ac6309c9bf702531f8b32c952547bc27e157527eaf67" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.418270 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "8983440f-2acf-40e0-a2fb-757413c0d0bb" (UID: "8983440f-2acf-40e0-a2fb-757413c0d0bb"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.433718 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement6c0f-account-delete-j5h7v" podStartSLOduration=4.433694942 podStartE2EDuration="4.433694942s" podCreationTimestamp="2025-12-09 11:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:58:38.240064365 +0000 UTC m=+1643.480822884" watchObservedRunningTime="2025-12-09 11:58:38.433694942 +0000 UTC m=+1643.674453461" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.449731 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi8c24-account-delete-ccctw"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.453336 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.453371 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8983440f-2acf-40e0-a2fb-757413c0d0bb-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.584713 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.960658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.988483 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.988781 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6b744fb85c-jt5h9" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-httpd" containerID="cri-o://b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90" gracePeriod=30 Dec 09 11:58:38 crc kubenswrapper[4770]: I1209 11:58:38.989393 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6b744fb85c-jt5h9" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-server" containerID="cri-o://fe68553e529f03896be808d62dfc73b3485de5dc39de9017e810c7baf0d4521c" gracePeriod=30 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.188204 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.219418 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.222967 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.239314 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_518c58bd-4a2f-439f-b5f5-0eda21434884/ovsdbserver-nb/0.log" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.241327 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.250468 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.259990 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.277935 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.301183 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.307544 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.307865 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308006 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2nw4\" (UniqueName: \"kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4\") pod \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308104 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b6jw\" (UniqueName: \"kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308256 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308402 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308568 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308622 4770 generic.go:334] "Generic (PLEG): container finished" podID="24be0696-be7e-46b9-aa90-d83d8abbd793" containerID="d42d0e251a4da94628ca8a8846c80d5cce60c71ad7836c652d111407395f41ec" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308732 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder4e4b-account-delete-5tttq" event={"ID":"24be0696-be7e-46b9-aa90-d83d8abbd793","Type":"ContainerDied","Data":"d42d0e251a4da94628ca8a8846c80d5cce60c71ad7836c652d111407395f41ec"} Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.308816 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.309422 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.309608 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.309724 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.309845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpqzv\" (UniqueName: \"kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv79v\" (UniqueName: \"kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310344 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310441 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310536 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310636 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310716 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310808 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.310936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom\") pod \"a95d5818-27fc-440a-b086-54eefa13d3ab\" (UID: \"a95d5818-27fc-440a-b086-54eefa13d3ab\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311270 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntppd\" (UniqueName: \"kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir\") pod \"518c58bd-4a2f-439f-b5f5-0eda21434884\" (UID: \"518c58bd-4a2f-439f-b5f5-0eda21434884\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311690 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret\") pod \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config\") pod \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311850 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb\") pod \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\" (UID: \"de2cc3bf-23a6-4b3f-a1fa-b845aad71787\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.311954 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle\") pod \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\" (UID: \"f6188401-2bb4-4ddc-a097-6b99f99df9e8\") " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.323466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config" (OuterVolumeSpecName: "config") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.328815 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.329006 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts" (OuterVolumeSpecName: "scripts") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.329465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.351769 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6363bd5-219b-4fb9-9695-7924791e3262" containerID="6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.351938 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.366160 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.378373 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd" (OuterVolumeSpecName: "kube-api-access-ntppd") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "kube-api-access-ntppd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.380041 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4" (OuterVolumeSpecName: "kube-api-access-p2nw4") pod "f6188401-2bb4-4ddc-a097-6b99f99df9e8" (UID: "f6188401-2bb4-4ddc-a097-6b99f99df9e8"). InnerVolumeSpecName "kube-api-access-p2nw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.380683 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3db1fbc4-24a1-4662-b146-064d33c9d5d6" path="/var/lib/kubelet/pods/3db1fbc4-24a1-4662-b146-064d33c9d5d6/volumes" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.381530 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" path="/var/lib/kubelet/pods/8983440f-2acf-40e0-a2fb-757413c0d0bb/volumes" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.382350 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5dfb341-2abb-4794-b559-1bee29b28016" path="/var/lib/kubelet/pods/b5dfb341-2abb-4794-b559-1bee29b28016/volumes" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.383721 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb132594-6dbe-452f-ab35-8337509cd95d" path="/var/lib/kubelet/pods/eb132594-6dbe-452f-ab35-8337509cd95d/volumes" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.385636 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.386475 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v" (OuterVolumeSpecName: "kube-api-access-bv79v") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "kube-api-access-bv79v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.386852 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.389035 4770 generic.go:334] "Generic (PLEG): container finished" podID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.395435 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv" (OuterVolumeSpecName: "kube-api-access-hpqzv") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "kube-api-access-hpqzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.396687 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw" (OuterVolumeSpecName: "kube-api-access-9b6jw") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "kube-api-access-9b6jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.397050 4770 generic.go:334] "Generic (PLEG): container finished" podID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerID="7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.400512 4770 generic.go:334] "Generic (PLEG): container finished" podID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerID="d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.400710 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.406681 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts" (OuterVolumeSpecName: "scripts") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.417424 4770 generic.go:334] "Generic (PLEG): container finished" podID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerID="b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422652 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b6jw\" (UniqueName: \"kubernetes.io/projected/a95d5818-27fc-440a-b086-54eefa13d3ab-kube-api-access-9b6jw\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422691 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2nw4\" (UniqueName: \"kubernetes.io/projected/f6188401-2bb4-4ddc-a097-6b99f99df9e8-kube-api-access-p2nw4\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422700 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422711 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpqzv\" (UniqueName: \"kubernetes.io/projected/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-kube-api-access-hpqzv\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422731 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv79v\" (UniqueName: \"kubernetes.io/projected/d6363bd5-219b-4fb9-9695-7924791e3262-kube-api-access-bv79v\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422747 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422759 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518c58bd-4a2f-439f-b5f5-0eda21434884-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422771 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a95d5818-27fc-440a-b086-54eefa13d3ab-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422782 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422815 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422831 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntppd\" (UniqueName: \"kubernetes.io/projected/518c58bd-4a2f-439f-b5f5-0eda21434884-kube-api-access-ntppd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.422843 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:39 crc kubenswrapper[4770]: E1209 11:58:39.581437 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:39 crc kubenswrapper[4770]: E1209 11:58:39.584472 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:39 crc kubenswrapper[4770]: E1209 11:58:39.589285 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:39 crc kubenswrapper[4770]: E1209 11:58:39.589370 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.596022 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.600146 4770 generic.go:334] "Generic (PLEG): container finished" podID="43a8a640-d610-4726-b2c6-84e99621d820" containerID="b72abc12a6683d231d24f05fd066d588593a15880270b01d02054a0969cc24af" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.604473 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_518c58bd-4a2f-439f-b5f5-0eda21434884/ovsdbserver-nb/0.log" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.609847 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.613567 4770 generic.go:334] "Generic (PLEG): container finished" podID="234ed53e-6da0-4fb5-8990-c4af4f500af8" containerID="245450190d772ae583ccda78ae92d3463c2252d79f297de71d0637f77ad6613f" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.625281 4770 generic.go:334] "Generic (PLEG): container finished" podID="846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" containerID="7aeca38e4fe31f32f1d0868743e6f73a07261e01e2d06fdcd6544df45818681d" exitCode=0 Dec 09 11:58:39 crc kubenswrapper[4770]: I1209 11:58:39.625394 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:39 crc kubenswrapper[4770]: E1209 11:58:39.771974 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb1b43fc_8d67_462b_a735_fcd38d19f7a8.slice/crio-b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb1b43fc_8d67_462b_a735_fcd38d19f7a8.slice/crio-conmon-b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.017595 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.089390 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.112504 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6b744fb85c-jt5h9" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.167:8080/healthcheck\": dial tcp 10.217.0.167:8080: connect: connection refused" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.113507 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6b744fb85c-jt5h9" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.167:8080/healthcheck\": dial tcp 10.217.0.167:8080: connect: connection refused" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.133601 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.195278 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.221778 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.225252 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.225407 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.227244 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.227300 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.236244 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.248523 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.248629 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.309027 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.309471 4770 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.309548 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:44.309529694 +0000 UTC m=+1649.550288223 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : configmap "openstack-cell1-scripts" not found Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.354763 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.478734 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") pod \"novacell17adb-account-delete-rtcfc\" (UID: \"0fccfd7f-e369-460c-86b6-16add156ef57\") " pod="openstack/novacell17adb-account-delete-rtcfc" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.479254 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.501152 4770 projected.go:194] Error preparing data for projected volume kube-api-access-5dpw7 for pod openstack/novacell17adb-account-delete-rtcfc: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.501235 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7 podName:0fccfd7f-e369-460c-86b6-16add156ef57 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:44.501211671 +0000 UTC m=+1649.741970190 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-5dpw7" (UniqueName: "kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7") pod "novacell17adb-account-delete-rtcfc" (UID: "0fccfd7f-e369-460c-86b6-16add156ef57") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.520968 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.537612 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.541694 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 11:58:40 crc kubenswrapper[4770]: E1209 11:58:40.541748 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerName="nova-scheduler-scheduler" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.565657 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" probeResult="failure" output=< Dec 09 11:58:40 crc kubenswrapper[4770]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Dec 09 11:58:40 crc kubenswrapper[4770]: > Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.586125 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:33668->10.217.0.201:8775: read: connection reset by peer" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.586254 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:33676->10.217.0.201:8775: read: connection reset by peer" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.672467 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:55870->10.217.0.159:9311: read: connection reset by peer" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.675287 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:55866->10.217.0.159:9311: read: connection reset by peer" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.758471 4770 generic.go:334] "Generic (PLEG): container finished" podID="a7a5dd25-a6fa-4afa-b107-453682e98cbb" containerID="66009135b64a083604820eb9655df507a1476d6aad063c1eff80d8350570d286" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.764023 4770 generic.go:334] "Generic (PLEG): container finished" podID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerID="9a473cbcc51156e661f62a9cb155beb880dd96bd5e5e0ee00faf14c8c0beeb93" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.768824 4770 generic.go:334] "Generic (PLEG): container finished" podID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerID="603f5bb321d03410442e30b0c6f4a276f62b4480a2b95aa1ef1c48a81bca45e8" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.792518 4770 generic.go:334] "Generic (PLEG): container finished" podID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerID="b505b07675385a501eefd16b47033f1d0cdb0b120345d67e47e86e33ef0f2548" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.795409 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.799732 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.802331 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.808948 4770 generic.go:334] "Generic (PLEG): container finished" podID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerID="eebed45a512ca372586490859e850a999b9d581d64647896a65dfb070e1af5fb" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.820194 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc74c6b9-5938-417d-bd71-062ecc932ced" containerID="e4009481f228df01d9b43e603422c8d7f751e18820ed452cc044d226f068d5c1" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.839429 4770 generic.go:334] "Generic (PLEG): container finished" podID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerID="28cd7242a9c9640679276453e171ec22b26b2720ffe571f2f85b629ee46a5e06" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.957806 4770 generic.go:334] "Generic (PLEG): container finished" podID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerID="2946bef94b26e9eb5227b621b0591908bb8f5db0d79016f7e9097a651629b7fc" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.960364 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.972517 4770 generic.go:334] "Generic (PLEG): container finished" podID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerID="fe68553e529f03896be808d62dfc73b3485de5dc39de9017e810c7baf0d4521c" exitCode=0 Dec 09 11:58:40 crc kubenswrapper[4770]: I1209 11:58:40.991754 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.080663 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.121580 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f6188401-2bb4-4ddc-a097-6b99f99df9e8" (UID: "f6188401-2bb4-4ddc-a097-6b99f99df9e8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.145113 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.179581 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.184312 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.185183 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.185227 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.191107 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell017c1-account-delete-xwp6r" podStartSLOduration=6.191081847 podStartE2EDuration="6.191081847s" podCreationTimestamp="2025-12-09 11:58:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:58:41.057887305 +0000 UTC m=+1646.298645824" watchObservedRunningTime="2025-12-09 11:58:41.191081847 +0000 UTC m=+1646.431840366" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.231456 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.264418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.289710 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.289754 4770 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.304093 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6188401-2bb4-4ddc-a097-6b99f99df9e8" (UID: "f6188401-2bb4-4ddc-a097-6b99f99df9e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.540660 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.540665 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.540787 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data podName:b596f863-b8d8-4fb4-93e3-45bdce78f1b1 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:49.540757576 +0000 UTC m=+1654.781516145 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data") pod "rabbitmq-cell1-server-0" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1") : configmap "rabbitmq-cell1-config-data" not found Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.591653 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.614244 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data" (OuterVolumeSpecName: "config-data") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.639516 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config" (OuterVolumeSpecName: "config") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.639643 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f6188401-2bb4-4ddc-a097-6b99f99df9e8" (UID: "f6188401-2bb4-4ddc-a097-6b99f99df9e8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.642092 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.643523 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") pod \"d6363bd5-219b-4fb9-9695-7924791e3262\" (UID: \"d6363bd5-219b-4fb9-9695-7924791e3262\") " Dec 09 11:58:41 crc kubenswrapper[4770]: W1209 11:58:41.643957 4770 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d6363bd5-219b-4fb9-9695-7924791e3262/volumes/kubernetes.io~secret/nova-novncproxy-tls-certs Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.643973 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "d6363bd5-219b-4fb9-9695-7924791e3262" (UID: "d6363bd5-219b-4fb9-9695-7924791e3262"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.644690 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.644710 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.644719 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.644728 4770 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6363bd5-219b-4fb9-9695-7924791e3262-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.644790 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f6188401-2bb4-4ddc-a097-6b99f99df9e8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.694194 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.706764 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.712040 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.712116 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.724927 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "518c58bd-4a2f-439f-b5f5-0eda21434884" (UID: "518c58bd-4a2f-439f-b5f5-0eda21434884"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.748303 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de2cc3bf-23a6-4b3f-a1fa-b845aad71787" (UID: "de2cc3bf-23a6-4b3f-a1fa-b845aad71787"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.766080 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/518c58bd-4a2f-439f-b5f5-0eda21434884-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.766121 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cc3bf-23a6-4b3f-a1fa-b845aad71787-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.806281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data" (OuterVolumeSpecName: "config-data") pod "a95d5818-27fc-440a-b086-54eefa13d3ab" (UID: "a95d5818-27fc-440a-b086-54eefa13d3ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.868699 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a95d5818-27fc-440a-b086-54eefa13d3ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.930588 4770 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.591s" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.930631 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6363bd5-219b-4fb9-9695-7924791e3262","Type":"ContainerDied","Data":"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.930679 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-jfg8z"] Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.941166 4770 scope.go:117] "RemoveContainer" containerID="6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.970058 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cabab723-8add-4dd0-b0cc-551f30a039d3" path="/var/lib/kubelet/pods/cabab723-8add-4dd0-b0cc-551f30a039d3/volumes" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.971486 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6188401-2bb4-4ddc-a097-6b99f99df9e8" path="/var/lib/kubelet/pods/f6188401-2bb4-4ddc-a097-6b99f99df9e8/volumes" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.976627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6363bd5-219b-4fb9-9695-7924791e3262","Type":"ContainerDied","Data":"d867aa5ccc5b0a777b2016533d93547869ee64ba952246a63fbcd7729967edc4"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.977634 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" event={"ID":"de2cc3bf-23a6-4b3f-a1fa-b845aad71787","Type":"ContainerDied","Data":"73989f64741448fd7cfbe906449b5e2863fedf1adec7d51f6605c5762a8d2648"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.977740 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8c24-account-delete-ccctw" event={"ID":"fc74c6b9-5938-417d-bd71-062ecc932ced","Type":"ContainerStarted","Data":"72fd7c31984f68a4fded3b1c5de09a4db94f31c2368d8ba02e5c36a1e8c8906c"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.977860 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell017c1-account-delete-xwp6r" event={"ID":"b1b663cc-0534-426b-ac89-cdd56dba00a5","Type":"ContainerStarted","Data":"8b0fddead96de7c1d1b699d68346b00f6dde46ed8b380b6bc1fbb2d94b32dd41"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.977963 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerDied","Data":"37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978107 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerDied","Data":"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978207 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerDied","Data":"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978292 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a95d5818-27fc-440a-b086-54eefa13d3ab","Type":"ContainerDied","Data":"645ea8faf00828041c7d4a7dc07574facddded447ba73ced65e32861022a7189"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978376 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance078e-account-delete-lwfnm" event={"ID":"a7a5dd25-a6fa-4afa-b107-453682e98cbb","Type":"ContainerStarted","Data":"af4d30672329911ea05faade3dfc7b0d08066f021327fafe05cbbc0ef8c9190a"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978470 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerDied","Data":"b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.978558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement6c0f-account-delete-j5h7v" event={"ID":"43a8a640-d610-4726-b2c6-84e99621d820","Type":"ContainerDied","Data":"b72abc12a6683d231d24f05fd066d588593a15880270b01d02054a0969cc24af"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979042 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"518c58bd-4a2f-439f-b5f5-0eda21434884","Type":"ContainerDied","Data":"80f870d5a6fc083f41c08a37e38b9c189c2fffdeffd28ec9149331a846b31a2a"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979139 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2ce4-account-delete-z5lmd" event={"ID":"234ed53e-6da0-4fb5-8990-c4af4f500af8","Type":"ContainerDied","Data":"245450190d772ae583ccda78ae92d3463c2252d79f297de71d0637f77ad6613f"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979226 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanaaaf-account-delete-x5lgz" event={"ID":"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb","Type":"ContainerDied","Data":"7aeca38e4fe31f32f1d0868743e6f73a07261e01e2d06fdcd6544df45818681d"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979324 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance078e-account-delete-lwfnm" event={"ID":"a7a5dd25-a6fa-4afa-b107-453682e98cbb","Type":"ContainerDied","Data":"66009135b64a083604820eb9655df507a1476d6aad063c1eff80d8350570d286"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerDied","Data":"9a473cbcc51156e661f62a9cb155beb880dd96bd5e5e0ee00faf14c8c0beeb93"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979491 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerDied","Data":"603f5bb321d03410442e30b0c6f4a276f62b4480a2b95aa1ef1c48a81bca45e8"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.979576 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerDied","Data":"b505b07675385a501eefd16b47033f1d0cdb0b120345d67e47e86e33ef0f2548"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.981862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-574ccd49f8-67m27" event={"ID":"356f02ca-284f-4ad4-9709-a9b6fba60b88","Type":"ContainerDied","Data":"26689815601d539a1e5a64d638a9b7e9d573190bd6a436f8545ce167279468c4"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.982063 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26689815601d539a1e5a64d638a9b7e9d573190bd6a436f8545ce167279468c4" Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.982168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerDied","Data":"eebed45a512ca372586490859e850a999b9d581d64647896a65dfb070e1af5fb"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.982256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8c24-account-delete-ccctw" event={"ID":"fc74c6b9-5938-417d-bd71-062ecc932ced","Type":"ContainerDied","Data":"e4009481f228df01d9b43e603422c8d7f751e18820ed452cc044d226f068d5c1"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.982336 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerDied","Data":"28cd7242a9c9640679276453e171ec22b26b2720ffe571f2f85b629ee46a5e06"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.982416 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerDied","Data":"2946bef94b26e9eb5227b621b0591908bb8f5db0d79016f7e9097a651629b7fc"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.983006 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerDied","Data":"fe68553e529f03896be808d62dfc73b3485de5dc39de9017e810c7baf0d4521c"} Dec 09 11:58:41 crc kubenswrapper[4770]: I1209 11:58:41.983187 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell017c1-account-delete-xwp6r" event={"ID":"b1b663cc-0534-426b-ac89-cdd56dba00a5","Type":"ContainerStarted","Data":"f7d03dfb52ad2f0872647ee11eede8299ffd7bddfcea9b460d3aa7f5fd2f65b2"} Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.987067 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 09 11:58:41 crc kubenswrapper[4770]: E1209 11:58:41.987156 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data podName:03fef9de-3f78-48b3-9079-9dc87184f803 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:49.987133451 +0000 UTC m=+1655.227891970 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data") pod "rabbitmq-server-0" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803") : configmap "rabbitmq-config-data" not found Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.036103 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerID="0193e4873f3244cab191ab4be299640f973319dfcf035f612a78cdfcd54c17a7" exitCode=0 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.036209 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerDied","Data":"0193e4873f3244cab191ab4be299640f973319dfcf035f612a78cdfcd54c17a7"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.057819 4770 generic.go:334] "Generic (PLEG): container finished" podID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerID="9b65050fb96dad2b68de130b2e74efdc4a249da8425436217ee12e2727d54d98" exitCode=0 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.057917 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerDied","Data":"9b65050fb96dad2b68de130b2e74efdc4a249da8425436217ee12e2727d54d98"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.068164 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c98e9f23-cf17-402d-8473-f10ba0c63a1d","Type":"ContainerDied","Data":"859d1a8633f3d060ab714d46ef835e7532895cb15b70431606765584fe929fbf"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.068217 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="859d1a8633f3d060ab714d46ef835e7532895cb15b70431606765584fe929fbf" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.078047 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc923d25-3beb-45cd-a504-0ebb6391a9f4","Type":"ContainerDied","Data":"7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.078129 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7613b93b658f14ddf13e1b0744437f1b70e2350adcef55629db689c1434617a8" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.083872 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1c0a1baa-962d-4dfc-891e-3563feff00bf","Type":"ContainerDied","Data":"714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.083949 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="714948f0c4af0e3cc4f594dba4167d05bc379382eed23a296b73c703e30efa74" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.086018 4770 generic.go:334] "Generic (PLEG): container finished" podID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerID="2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" exitCode=0 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.086074 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cead551d-3523-499a-a1b4-ebce9c97e59e","Type":"ContainerDied","Data":"2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.102538 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb","Type":"ContainerDied","Data":"8351da7f332ac8f54987f847f02c25d2625de68f0947f11791e8e48f707cdbd6"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.102579 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8351da7f332ac8f54987f847f02c25d2625de68f0947f11791e8e48f707cdbd6" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.118730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b","Type":"ContainerDied","Data":"a52884cbb85f13029b50a0ff7af99099b7aabcafb4a808f4b3590340e449e81b"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.118790 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a52884cbb85f13029b50a0ff7af99099b7aabcafb4a808f4b3590340e449e81b" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.127775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b744fb85c-jt5h9" event={"ID":"eb1b43fc-8d67-462b-a735-fcd38d19f7a8","Type":"ContainerDied","Data":"01e26f55a994df9fdd65a6df86e3575b0a563a3c1822511ef020f6f8c9b66a18"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.127816 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01e26f55a994df9fdd65a6df86e3575b0a563a3c1822511ef020f6f8c9b66a18" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.133100 4770 generic.go:334] "Generic (PLEG): container finished" podID="b1b663cc-0534-426b-ac89-cdd56dba00a5" containerID="f7d03dfb52ad2f0872647ee11eede8299ffd7bddfcea9b460d3aa7f5fd2f65b2" exitCode=0 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.133822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell017c1-account-delete-xwp6r" event={"ID":"b1b663cc-0534-426b-ac89-cdd56dba00a5","Type":"ContainerDied","Data":"f7d03dfb52ad2f0872647ee11eede8299ffd7bddfcea9b460d3aa7f5fd2f65b2"} Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.254531 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.259679 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="sg-core" containerID="cri-o://b85b05cd17651d9b3bd353aa522c2eeb740da829a60de5831bcd5ea98ce4117d" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.259866 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="proxy-httpd" containerID="cri-o://d3cff712577cb5ca4995f819406108099e3cabf583e0b9fe83d7a120676f4548" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.259968 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-notification-agent" containerID="cri-o://bd1a0426233ccd1b1386234cfd700d23db549b81974daa36103a0f808577669b" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.267562 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-central-agent" containerID="cri-o://07ea87393e7d08e9bbfe345d3786c9fd9565d8d1f4171600282eccd2c3cc50f2" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.319857 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.320210 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerName="kube-state-metrics" containerID="cri-o://218393e2da0bbfb5410ec4757c8cb027246a3f4ac476769670c88ceb23a30731" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.516475 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.539392 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.539598 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" containerName="memcached" containerID="cri-o://a7dc5c4c948b7e59408c0e3cf9bcfbf9cd1717522c72afefd9d893df384cec6a" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.553658 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.634491 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jzfqf"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.677811 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jh8bg"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.711586 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jzfqf"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.742767 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jh8bg"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.766727 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.767655 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5b4585bb4d-2j2md" podUID="b306554d-ccb5-4db1-8462-fef6aba58f57" containerName="keystone-api" containerID="cri-o://97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919" gracePeriod=30 Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.796885 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone9545-account-delete-8lfqw"] Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797734 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="init" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797758 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="init" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797784 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="ovsdbserver-sb" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797792 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="ovsdbserver-sb" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797802 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797808 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797834 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="cinder-scheduler" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797839 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="cinder-scheduler" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797852 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797859 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797866 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="ovsdbserver-nb" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797872 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="ovsdbserver-nb" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797887 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cabab723-8add-4dd0-b0cc-551f30a039d3" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797892 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cabab723-8add-4dd0-b0cc-551f30a039d3" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797928 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797940 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797952 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="probe" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797959 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="probe" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.797971 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="dnsmasq-dns" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.797977 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="dnsmasq-dns" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798193 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="cinder-scheduler" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798204 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" containerName="probe" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798215 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cabab723-8add-4dd0-b0cc-551f30a039d3" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798227 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="ovsdbserver-sb" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798240 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798251 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" containerName="ovsdbserver-nb" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798260 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798401 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="dnsmasq-dns" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.798412 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8983440f-2acf-40e0-a2fb-757413c0d0bb" containerName="openstack-network-exporter" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.799503 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.817736 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-867cd545c7-gzq6m" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.195:5353: i/o timeout" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.845134 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.857822 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone9545-account-delete-8lfqw"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.870929 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-fmjv6"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.876891 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.877025 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.884963 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-fmjv6"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.896017 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9545-account-create-update-4vn85"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.904619 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9545-account-create-update-4vn85"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.927843 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone9545-account-delete-8lfqw"] Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.984668 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:42 crc kubenswrapper[4770]: I1209 11:58:42.985359 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.986056 4770 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.986142 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:43.486126677 +0000 UTC m=+1648.726885186 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : configmap "openstack-scripts" not found Dec 09 11:58:42 crc kubenswrapper[4770]: E1209 11:58:42.998381 4770 projected.go:194] Error preparing data for projected volume kube-api-access-xfqwf for pod openstack/keystone9545-account-delete-8lfqw: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.002205 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:43.502168985 +0000 UTC m=+1648.742927504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xfqwf" (UniqueName: "kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.011089 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.187:8081/readyz\": dial tcp 10.217.0.187:8081: connect: connection refused" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.052210 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="galera" containerID="cri-o://79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" gracePeriod=30 Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.098190 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.100179 4770 scope.go:117] "RemoveContainer" containerID="6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.104191 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130\": container with ID starting with 6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130 not found: ID does not exist" containerID="6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.104231 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130"} err="failed to get container status \"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130\": rpc error: code = NotFound desc = could not find container \"6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130\": container with ID starting with 6c87b2ebff7e5c24365653dc8777b7e43006c983440be6beea08a9781abdb130 not found: ID does not exist" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.104277 4770 scope.go:117] "RemoveContainer" containerID="43ebc26e2c2fa580b48c4844f5bfe5b3ba67d22ccd82e4aaaa68755ef5a9c123" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.126926 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell17adb-account-delete-rtcfc"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.153097 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell17adb-account-delete-rtcfc"] Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.190486 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194381 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194460 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194514 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194551 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194589 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdjqb\" (UniqueName: \"kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.194882 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts\") pod \"356f02ca-284f-4ad4-9709-a9b6fba60b88\" (UID: \"356f02ca-284f-4ad4-9709-a9b6fba60b88\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.195582 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs" (OuterVolumeSpecName: "logs") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.196002 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356f02ca-284f-4ad4-9709-a9b6fba60b88-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.206633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanaaaf-account-delete-x5lgz" event={"ID":"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb","Type":"ContainerDied","Data":"9e890bcfd1a493e1a8d379b97b5e7d8b67b9f9c84627c7cfdefc69f24605ef06"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.206675 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e890bcfd1a493e1a8d379b97b5e7d8b67b9f9c84627c7cfdefc69f24605ef06" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.215854 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.219201 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb" (OuterVolumeSpecName: "kube-api-access-jdjqb") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "kube-api-access-jdjqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.225454 4770 generic.go:334] "Generic (PLEG): container finished" podID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerID="218393e2da0bbfb5410ec4757c8cb027246a3f4ac476769670c88ceb23a30731" exitCode=2 Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.225542 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b","Type":"ContainerDied","Data":"218393e2da0bbfb5410ec4757c8cb027246a3f4ac476769670c88ceb23a30731"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.228095 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts" (OuterVolumeSpecName: "scripts") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.232754 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement6c0f-account-delete-j5h7v" event={"ID":"43a8a640-d610-4726-b2c6-84e99621d820","Type":"ContainerDied","Data":"3442fb69865baa3d225d215aefb85c5bdf397a7a1f6e869fd463140e7857c2fd"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.232805 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3442fb69865baa3d225d215aefb85c5bdf397a7a1f6e869fd463140e7857c2fd" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.238839 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.238936 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="galera" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.247745 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8c24-account-delete-ccctw" event={"ID":"fc74c6b9-5938-417d-bd71-062ecc932ced","Type":"ContainerDied","Data":"72fd7c31984f68a4fded3b1c5de09a4db94f31c2368d8ba02e5c36a1e8c8906c"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.247791 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72fd7c31984f68a4fded3b1c5de09a4db94f31c2368d8ba02e5c36a1e8c8906c" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.289813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.300378 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fccfd7f-e369-460c-86b6-16add156ef57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.300455 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dpw7\" (UniqueName: \"kubernetes.io/projected/0fccfd7f-e369-460c-86b6-16add156ef57-kube-api-access-5dpw7\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.300471 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.300487 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.300499 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdjqb\" (UniqueName: \"kubernetes.io/projected/356f02ca-284f-4ad4-9709-a9b6fba60b88-kube-api-access-jdjqb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.305944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9d1c6e5-c685-442c-94bb-a6f650adb837","Type":"ContainerDied","Data":"e663a780d9b06b4e39849bb7b0c4661407033e945777e41c7fb8eca7c8e320f3"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.305997 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e663a780d9b06b4e39849bb7b0c4661407033e945777e41c7fb8eca7c8e320f3" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.308626 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.310381 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance078e-account-delete-lwfnm" event={"ID":"a7a5dd25-a6fa-4afa-b107-453682e98cbb","Type":"ContainerDied","Data":"af4d30672329911ea05faade3dfc7b0d08066f021327fafe05cbbc0ef8c9190a"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.311355 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af4d30672329911ea05faade3dfc7b0d08066f021327fafe05cbbc0ef8c9190a" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.310649 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.314375 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2ce4-account-delete-z5lmd" event={"ID":"234ed53e-6da0-4fb5-8990-c4af4f500af8","Type":"ContainerDied","Data":"3eec329e3fd54ef56513a36e5cdc372b32f3a824e2bb2c5cffd5857e1d917987"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.314416 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eec329e3fd54ef56513a36e5cdc372b32f3a824e2bb2c5cffd5857e1d917987" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.321075 4770 scope.go:117] "RemoveContainer" containerID="06e985e24579ff9c413e840e9b0638c771aa22054a9c57c3468b29de99f1d1da" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.321977 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder4e4b-account-delete-5tttq" event={"ID":"24be0696-be7e-46b9-aa90-d83d8abbd793","Type":"ContainerDied","Data":"dc4d3f4170827dac7881677505a31fbd90cfd992ad89a750ab423222c74bfc3f"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.322067 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc4d3f4170827dac7881677505a31fbd90cfd992ad89a750ab423222c74bfc3f" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.352716 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.364972 4770 scope.go:117] "RemoveContainer" containerID="7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.384739 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fccfd7f-e369-460c-86b6-16add156ef57" path="/var/lib/kubelet/pods/0fccfd7f-e369-460c-86b6-16add156ef57/volumes" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.385467 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109b0cbe-45dc-416c-a4a2-972eda68baf2" path="/var/lib/kubelet/pods/109b0cbe-45dc-416c-a4a2-972eda68baf2/volumes" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.386768 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ff5c57-41ce-43a3-b65e-7d52dd45ee3f" path="/var/lib/kubelet/pods/18ff5c57-41ce-43a3-b65e-7d52dd45ee3f/volumes" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.387850 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5dc7848-9e9b-43e2-8800-b7c6678c1eab" path="/var/lib/kubelet/pods/a5dc7848-9e9b-43e2-8800-b7c6678c1eab/volumes" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.389584 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647" path="/var/lib/kubelet/pods/b8299ef4-cea1-4ddc-b9a7-dd61fb6b9647/volumes" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650305 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650358 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650405 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650453 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650495 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp6jk\" (UniqueName: \"kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650538 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650582 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650686 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650739 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzpcn\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650820 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.650848 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651001 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651063 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651090 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651192 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs\") pod \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\" (UID: \"dc923d25-3beb-45cd-a504-0ebb6391a9f4\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651230 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbf9p\" (UniqueName: \"kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p\") pod \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\" (UID: \"f17f6cc7-50ef-45f1-9552-edc09f1e6ecb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651277 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle\") pod \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\" (UID: \"eb1b43fc-8d67-462b-a735-fcd38d19f7a8\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651957 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.651870 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.652125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.652296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.653018 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.653885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs" (OuterVolumeSpecName: "logs") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.653984 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.665671 4770 projected.go:194] Error preparing data for projected volume kube-api-access-xfqwf for pod openstack/keystone9545-account-delete-8lfqw: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.665735 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:44.665719544 +0000 UTC m=+1649.906478063 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-xfqwf" (UniqueName: "kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.666202 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.666304 4770 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.666406 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:44.666389351 +0000 UTC m=+1649.907147870 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : configmap "openstack-scripts" not found Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.666946 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.668618 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerID="d3cff712577cb5ca4995f819406108099e3cabf583e0b9fe83d7a120676f4548" exitCode=0 Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.668728 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerID="b85b05cd17651d9b3bd353aa522c2eeb740da829a60de5831bcd5ea98ce4117d" exitCode=2 Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.668889 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-574ccd49f8-67m27" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.679823 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.689540 4770 scope.go:117] "RemoveContainer" containerID="d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.695109 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk" (OuterVolumeSpecName: "kube-api-access-rp6jk") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "kube-api-access-rp6jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696261 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5674985874-bhmwx" event={"ID":"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3","Type":"ContainerDied","Data":"24d4877fa872b6053bef5de4f870b65764d0ce5e9ed33b03b1a6a39665590b34"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696309 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24d4877fa872b6053bef5de4f870b65764d0ce5e9ed33b03b1a6a39665590b34" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cead551d-3523-499a-a1b4-ebce9c97e59e","Type":"ContainerDied","Data":"94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696351 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94736e91baf6e847941598ffdbc8bb4bd84f2f9459568506b06ea2ac8a89cdf3" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696361 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerDied","Data":"d3cff712577cb5ca4995f819406108099e3cabf583e0b9fe83d7a120676f4548"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.696379 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerDied","Data":"b85b05cd17651d9b3bd353aa522c2eeb740da829a60de5831bcd5ea98ce4117d"} Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.706851 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p" (OuterVolumeSpecName: "kube-api-access-cbf9p") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "kube-api-access-cbf9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.706960 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn" (OuterVolumeSpecName: "kube-api-access-vzpcn") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "kube-api-access-vzpcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.707522 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.709154 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.729880 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.729962 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753757 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753829 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753881 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.753973 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.754241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.754879 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9mq\" (UniqueName: \"kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.754973 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755176 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755200 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755314 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755391 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z92n\" (UniqueName: \"kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755438 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755523 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle\") pod \"1c0a1baa-962d-4dfc-891e-3563feff00bf\" (UID: \"1c0a1baa-962d-4dfc-891e-3563feff00bf\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.755569 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\" (UID: \"9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756294 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc923d25-3beb-45cd-a504-0ebb6391a9f4-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756308 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbf9p\" (UniqueName: \"kubernetes.io/projected/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kube-api-access-cbf9p\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756318 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756328 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756337 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp6jk\" (UniqueName: \"kubernetes.io/projected/dc923d25-3beb-45cd-a504-0ebb6391a9f4-kube-api-access-rp6jk\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756345 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756369 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756380 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756389 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzpcn\" (UniqueName: \"kubernetes.io/projected/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-kube-api-access-vzpcn\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756402 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756414 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.756425 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.761059 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.762247 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.768201 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.771824 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs" (OuterVolumeSpecName: "logs") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.771969 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.773744 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs" (OuterVolumeSpecName: "logs") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.774082 4770 scope.go:117] "RemoveContainer" containerID="7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.774319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.775254 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.775338 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd\": container with ID starting with 7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd not found: ID does not exist" containerID="7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.775366 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd"} err="failed to get container status \"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd\": rpc error: code = NotFound desc = could not find container \"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd\": container with ID starting with 7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd not found: ID does not exist" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.775394 4770 scope.go:117] "RemoveContainer" containerID="d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" Dec 09 11:58:43 crc kubenswrapper[4770]: E1209 11:58:43.777976 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4\": container with ID starting with d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4 not found: ID does not exist" containerID="d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.778557 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4"} err="failed to get container status \"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4\": rpc error: code = NotFound desc = could not find container \"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4\": container with ID starting with d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4 not found: ID does not exist" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.778670 4770 scope.go:117] "RemoveContainer" containerID="7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.783813 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd"} err="failed to get container status \"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd\": rpc error: code = NotFound desc = could not find container \"7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd\": container with ID starting with 7ae4459cc636e711c948789a221b77e37645fb0a41ba8f5519038412dd01f7dd not found: ID does not exist" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.783856 4770 scope.go:117] "RemoveContainer" containerID="d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.790483 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4"} err="failed to get container status \"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4\": rpc error: code = NotFound desc = could not find container \"d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4\": container with ID starting with d6b22fa065f309d5792ef36565c28d59f55967368c2c777c545f5926034a3be4 not found: ID does not exist" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.790542 4770 scope.go:117] "RemoveContainer" containerID="faf973a0c6d5bcf16a241dc80a13699d03f0a874da0549eb1f5e4001556d13ae" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.798841 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.803985 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.814931 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq" (OuterVolumeSpecName: "kube-api-access-2w9mq") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "kube-api-access-2w9mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.818286 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.823294 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.826211 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.835244 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts" (OuterVolumeSpecName: "scripts") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.839350 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.840522 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts" (OuterVolumeSpecName: "scripts") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.840884 4770 scope.go:117] "RemoveContainer" containerID="3ca7cb6b136b6f5203dbd38211b8cbec6b683499790baf062a6e08bed8bcd01b" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.841104 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867cd545c7-gzq6m"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.851079 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n" (OuterVolumeSpecName: "kube-api-access-8z92n") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "kube-api-access-8z92n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859665 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvl48\" (UniqueName: \"kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq2xz\" (UniqueName: \"kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz\") pod \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859837 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle\") pod \"cead551d-3523-499a-a1b4-ebce9c97e59e\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859915 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hfdh\" (UniqueName: \"kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh\") pod \"cead551d-3523-499a-a1b4-ebce9c97e59e\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859945 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs\") pod \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.859980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data\") pod \"cead551d-3523-499a-a1b4-ebce9c97e59e\" (UID: \"cead551d-3523-499a-a1b4-ebce9c97e59e\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860041 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860078 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8486\" (UniqueName: \"kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486\") pod \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860110 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts\") pod \"43a8a640-d610-4726-b2c6-84e99621d820\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860174 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860201 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts\") pod \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\" (UID: \"846d1db4-8766-42c2-8fe4-3ed3ba6b80fb\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.860242 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs\") pod \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.861214 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43a8a640-d610-4726-b2c6-84e99621d820" (UID: "43a8a640-d610-4726-b2c6-84e99621d820"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.875228 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" (UID: "846d1db4-8766-42c2-8fe4-3ed3ba6b80fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891423 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql2j4\" (UniqueName: \"kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4\") pod \"43a8a640-d610-4726-b2c6-84e99621d820\" (UID: \"43a8a640-d610-4726-b2c6-84e99621d820\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.890293 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891537 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891612 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data\") pod \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891647 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891672 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle\") pod \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\" (UID: \"c98e9f23-cf17-402d-8473-f10ba0c63a1d\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891728 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.891786 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs\") pod \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\" (UID: \"3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3\") " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893678 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893701 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z92n\" (UniqueName: \"kubernetes.io/projected/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-kube-api-access-8z92n\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893715 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893741 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893751 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893760 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893768 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c0a1baa-962d-4dfc-891e-3563feff00bf-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893776 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893784 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893862 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9mq\" (UniqueName: \"kubernetes.io/projected/1c0a1baa-962d-4dfc-891e-3563feff00bf-kube-api-access-2w9mq\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893872 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0a1baa-962d-4dfc-891e-3563feff00bf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893888 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893909 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893918 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43a8a640-d610-4726-b2c6-84e99621d820-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893827 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs" (OuterVolumeSpecName: "logs") pod "c98e9f23-cf17-402d-8473-f10ba0c63a1d" (UID: "c98e9f23-cf17-402d-8473-f10ba0c63a1d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.893894 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs" (OuterVolumeSpecName: "logs") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.899193 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh" (OuterVolumeSpecName: "kube-api-access-7hfdh") pod "cead551d-3523-499a-a1b4-ebce9c97e59e" (UID: "cead551d-3523-499a-a1b4-ebce9c97e59e"). InnerVolumeSpecName "kube-api-access-7hfdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.901105 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.918172 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.930091 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.931613 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz" (OuterVolumeSpecName: "kube-api-access-zq2xz") pod "846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" (UID: "846d1db4-8766-42c2-8fe4-3ed3ba6b80fb"). InnerVolumeSpecName "kube-api-access-zq2xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.937690 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.938044 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486" (OuterVolumeSpecName: "kube-api-access-s8486") pod "c98e9f23-cf17-402d-8473-f10ba0c63a1d" (UID: "c98e9f23-cf17-402d-8473-f10ba0c63a1d"). InnerVolumeSpecName "kube-api-access-s8486". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.938082 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data" (OuterVolumeSpecName: "config-data") pod "356f02ca-284f-4ad4-9709-a9b6fba60b88" (UID: "356f02ca-284f-4ad4-9709-a9b6fba60b88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.945182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48" (OuterVolumeSpecName: "kube-api-access-lvl48") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "kube-api-access-lvl48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:43 crc kubenswrapper[4770]: I1209 11:58:43.956241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4" (OuterVolumeSpecName: "kube-api-access-ql2j4") pod "43a8a640-d610-4726-b2c6-84e99621d820" (UID: "43a8a640-d610-4726-b2c6-84e99621d820"). InnerVolumeSpecName "kube-api-access-ql2j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.017512 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042156 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql2j4\" (UniqueName: \"kubernetes.io/projected/43a8a640-d610-4726-b2c6-84e99621d820-kube-api-access-ql2j4\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042207 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042223 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356f02ca-284f-4ad4-9709-a9b6fba60b88-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042236 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvl48\" (UniqueName: \"kubernetes.io/projected/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-kube-api-access-lvl48\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042252 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq2xz\" (UniqueName: \"kubernetes.io/projected/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb-kube-api-access-zq2xz\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042265 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hfdh\" (UniqueName: \"kubernetes.io/projected/cead551d-3523-499a-a1b4-ebce9c97e59e-kube-api-access-7hfdh\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042277 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042291 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042303 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8486\" (UniqueName: \"kubernetes.io/projected/c98e9f23-cf17-402d-8473-f10ba0c63a1d-kube-api-access-s8486\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.042315 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c98e9f23-cf17-402d-8473-f10ba0c63a1d-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.159894 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cead551d-3523-499a-a1b4-ebce9c97e59e" (UID: "cead551d-3523-499a-a1b4-ebce9c97e59e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.250885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.261709 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.261747 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.270048 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.284350 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" (UID: "f17f6cc7-50ef-45f1-9552-edc09f1e6ecb"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.294895 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.334001 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-xx2q9"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.336502 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.344216 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-xx2q9"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.371412 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement6c0f-account-delete-j5h7v"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.373957 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.373997 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.380479 4770 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.380523 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.381596 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6c0f-account-create-update-sstlm"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.391548 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement6c0f-account-delete-j5h7v"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.403232 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6c0f-account-create-update-sstlm"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.448668 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.454104 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.457998 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.467507 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.469677 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.473135 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.473227 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.479684 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.483543 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.483582 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.483595 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.483607 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.485382 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c98e9f23-cf17-402d-8473-f10ba0c63a1d" (UID: "c98e9f23-cf17-402d-8473-f10ba0c63a1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.521254 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.539363 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data" (OuterVolumeSpecName: "config-data") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.547098 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.568608 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data" (OuterVolumeSpecName: "config-data") pod "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" (UID: "9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.589523 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.589559 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.589572 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.589581 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.589590 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.636106 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data" (OuterVolumeSpecName: "config-data") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.636957 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.637181 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c98e9f23-cf17-402d-8473-f10ba0c63a1d" (UID: "c98e9f23-cf17-402d-8473-f10ba0c63a1d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.641690 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data" (OuterVolumeSpecName: "config-data") pod "cead551d-3523-499a-a1b4-ebce9c97e59e" (UID: "cead551d-3523-499a-a1b4-ebce9c97e59e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.643122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.655946 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data" (OuterVolumeSpecName: "config-data") pod "dc923d25-3beb-45cd-a504-0ebb6391a9f4" (UID: "dc923d25-3beb-45cd-a504-0ebb6391a9f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.658628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.660124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" (UID: "3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.661030 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.663224 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c0a1baa-962d-4dfc-891e-3563feff00bf" (UID: "1c0a1baa-962d-4dfc-891e-3563feff00bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.665451 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lrtgq"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.677870 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data" (OuterVolumeSpecName: "config-data") pod "eb1b43fc-8d67-462b-a735-fcd38d19f7a8" (UID: "eb1b43fc-8d67-462b-a735-fcd38d19f7a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.678212 4770 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 09 11:58:44 crc kubenswrapper[4770]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-09T11:58:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 09 11:58:44 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 456 Alarm clock "$@" Dec 09 11:58:44 crc kubenswrapper[4770]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-xhnq4" message=< Dec 09 11:58:44 crc kubenswrapper[4770]: Exiting ovn-controller (1) [FAILED] Dec 09 11:58:44 crc kubenswrapper[4770]: Killing ovn-controller (1) [ OK ] Dec 09 11:58:44 crc kubenswrapper[4770]: Killing ovn-controller (1) with SIGKILL [ OK ] Dec 09 11:58:44 crc kubenswrapper[4770]: 2025-12-09T11:58:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 09 11:58:44 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 456 Alarm clock "$@" Dec 09 11:58:44 crc kubenswrapper[4770]: > Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.678249 4770 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 09 11:58:44 crc kubenswrapper[4770]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-09T11:58:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 09 11:58:44 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 456 Alarm clock "$@" Dec 09 11:58:44 crc kubenswrapper[4770]: > pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" containerID="cri-o://819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.678320 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" containerID="cri-o://819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" gracePeriod=21 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.683611 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data" (OuterVolumeSpecName: "config-data") pod "c98e9f23-cf17-402d-8473-f10ba0c63a1d" (UID: "c98e9f23-cf17-402d-8473-f10ba0c63a1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.706758 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707302 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707330 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707345 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc923d25-3beb-45cd-a504-0ebb6391a9f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707358 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707371 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707389 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707403 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707415 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b43fc-8d67-462b-a735-fcd38d19f7a8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707432 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c98e9f23-cf17-402d-8473-f10ba0c63a1d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707449 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cead551d-3523-499a-a1b4-ebce9c97e59e-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707483 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c0a1baa-962d-4dfc-891e-3563feff00bf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.707498 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.707606 4770 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.707689 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:46.70766846 +0000 UTC m=+1651.948426979 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : configmap "openstack-scripts" not found Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.714548 4770 projected.go:194] Error preparing data for projected volume kube-api-access-xfqwf for pod openstack/keystone9545-account-delete-8lfqw: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.714653 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:46.714620307 +0000 UTC m=+1651.955378826 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-xfqwf" (UniqueName: "kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.724982 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lrtgq"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.771160 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.794551 4770 generic.go:334] "Generic (PLEG): container finished" podID="9f7b544a-d135-44fc-9fda-3938369f661f" containerID="bf8ad42774200d477849ea268cd29b0c65b5c006ef52af1b6710676014a7e842" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.794934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerDied","Data":"bf8ad42774200d477849ea268cd29b0c65b5c006ef52af1b6710676014a7e842"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.831349 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.837862 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.838433 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b","Type":"ContainerDied","Data":"a9346dc03c35e03914390456553753b562c2a058abe52edb39c068776fc9b55b"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.838481 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9346dc03c35e03914390456553753b562c2a058abe52edb39c068776fc9b55b" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.839608 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.839737 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.839808 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.877049 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq4dr\" (UniqueName: \"kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.877124 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.877160 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.877192 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.877261 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data\") pod \"a9d1c6e5-c685-442c-94bb-a6f650adb837\" (UID: \"a9d1c6e5-c685-442c-94bb-a6f650adb837\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.878306 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs" (OuterVolumeSpecName: "logs") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.879196 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4e4b-account-create-update-shln7"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.879513 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.882156 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.888738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr" (OuterVolumeSpecName: "kube-api-access-sq4dr") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "kube-api-access-sq4dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.892939 4770 generic.go:334] "Generic (PLEG): container finished" podID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerID="f2d4dc50bf2b36a841745128ed7fc9d07db8df1ab56a2af6bc7a83543d7f633f" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.893045 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerDied","Data":"f2d4dc50bf2b36a841745128ed7fc9d07db8df1ab56a2af6bc7a83543d7f633f"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.893079 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b596f863-b8d8-4fb4-93e3-45bdce78f1b1","Type":"ContainerDied","Data":"98e3471866b7b3ea7ac43432961287c5a59b6b0f6249f5401b19933d817b43a7"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.893094 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98e3471866b7b3ea7ac43432961287c5a59b6b0f6249f5401b19933d817b43a7" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.899881 4770 scope.go:117] "RemoveContainer" containerID="5525f70a6a808175e62c919c6f6e56ec54293e10c519fefa5b6d1b5a45deea6e" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.908020 4770 generic.go:334] "Generic (PLEG): container finished" podID="df40ca35-a10e-4571-b460-f7cd465070d8" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.908140 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.908169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"df40ca35-a10e-4571-b460-f7cd465070d8","Type":"ContainerDied","Data":"5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.914683 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts" (OuterVolumeSpecName: "scripts") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.924051 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4e4b-account-create-update-shln7"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.929498 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-j4q62"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.940249 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerID="07ea87393e7d08e9bbfe345d3786c9fd9565d8d1f4171600282eccd2c3cc50f2" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.940368 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-j4q62"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.940399 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerDied","Data":"07ea87393e7d08e9bbfe345d3786c9fd9565d8d1f4171600282eccd2c3cc50f2"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.953619 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell017c1-account-delete-xwp6r" event={"ID":"b1b663cc-0534-426b-ac89-cdd56dba00a5","Type":"ContainerDied","Data":"8b0fddead96de7c1d1b699d68346b00f6dde46ed8b380b6bc1fbb2d94b32dd41"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.953658 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b0fddead96de7c1d1b699d68346b00f6dde46ed8b380b6bc1fbb2d94b32dd41" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.963768 4770 generic.go:334] "Generic (PLEG): container finished" podID="568ef04e-032c-40f4-97b9-b21483281a4f" containerID="3390dfb8cbb1f0c6bd1a75c816ce715ca36a3164d095376f2127e25ddb826b67" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.963926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerDied","Data":"3390dfb8cbb1f0c6bd1a75c816ce715ca36a3164d095376f2127e25ddb826b67"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.965134 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.968603 4770 generic.go:334] "Generic (PLEG): container finished" podID="03fef9de-3f78-48b3-9079-9dc87184f803" containerID="1980ac00c990360135915c858d0f125e136b46aa153c5a4d5cf4019dd84f4355" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.968789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerDied","Data":"1980ac00c990360135915c858d0f125e136b46aa153c5a4d5cf4019dd84f4355"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.968818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03fef9de-3f78-48b3-9079-9dc87184f803","Type":"ContainerDied","Data":"372b4b00e2cc2abd1ce34b82c880ec351754b64a39d8320cda277799057ff294"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.968830 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="372b4b00e2cc2abd1ce34b82c880ec351754b64a39d8320cda277799057ff294" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.972407 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-aaaf-account-create-update-bv8tm"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.980017 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh864\" (UniqueName: \"kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864\") pod \"234ed53e-6da0-4fb5-8990-c4af4f500af8\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.982045 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcdd5\" (UniqueName: \"kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5\") pod \"24be0696-be7e-46b9-aa90-d83d8abbd793\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.982090 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts\") pod \"24be0696-be7e-46b9-aa90-d83d8abbd793\" (UID: \"24be0696-be7e-46b9-aa90-d83d8abbd793\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.982136 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts\") pod \"234ed53e-6da0-4fb5-8990-c4af4f500af8\" (UID: \"234ed53e-6da0-4fb5-8990-c4af4f500af8\") " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.982311 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanaaaf-account-delete-x5lgz"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.985812 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5" (OuterVolumeSpecName: "kube-api-access-bcdd5") pod "24be0696-be7e-46b9-aa90-d83d8abbd793" (UID: "24be0696-be7e-46b9-aa90-d83d8abbd793"). InnerVolumeSpecName "kube-api-access-bcdd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.986457 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24be0696-be7e-46b9-aa90-d83d8abbd793" (UID: "24be0696-be7e-46b9-aa90-d83d8abbd793"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.987064 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "234ed53e-6da0-4fb5-8990-c4af4f500af8" (UID: "234ed53e-6da0-4fb5-8990-c4af4f500af8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: E1209 11:58:44.992252 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02 is running failed: container process not found" containerID="819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.993117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864" (OuterVolumeSpecName: "kube-api-access-xh864") pod "234ed53e-6da0-4fb5-8990-c4af4f500af8" (UID: "234ed53e-6da0-4fb5-8990-c4af4f500af8"). InnerVolumeSpecName "kube-api-access-xh864". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.993181 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-aaaf-account-create-update-bv8tm"] Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994462 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcdd5\" (UniqueName: \"kubernetes.io/projected/24be0696-be7e-46b9-aa90-d83d8abbd793-kube-api-access-bcdd5\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994498 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24be0696-be7e-46b9-aa90-d83d8abbd793-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994510 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234ed53e-6da0-4fb5-8990-c4af4f500af8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994546 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994564 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994574 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh864\" (UniqueName: \"kubernetes.io/projected/234ed53e-6da0-4fb5-8990-c4af4f500af8-kube-api-access-xh864\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994584 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq4dr\" (UniqueName: \"kubernetes.io/projected/a9d1c6e5-c685-442c-94bb-a6f650adb837-kube-api-access-sq4dr\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994595 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994613 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.994623 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9d1c6e5-c685-442c-94bb-a6f650adb837-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.995224 4770 generic.go:334] "Generic (PLEG): container finished" podID="f352a11c-1887-4736-98cb-4eccf3086e97" containerID="a7dc5c4c948b7e59408c0e3cf9bcfbf9cd1717522c72afefd9d893df384cec6a" exitCode=0 Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.995371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f352a11c-1887-4736-98cb-4eccf3086e97","Type":"ContainerDied","Data":"a7dc5c4c948b7e59408c0e3cf9bcfbf9cd1717522c72afefd9d893df384cec6a"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.995407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f352a11c-1887-4736-98cb-4eccf3086e97","Type":"ContainerDied","Data":"7a853b3079ae327a5e2751c66ecf028c4a7e55fd1f1c6894af589afedfaf48e0"} Dec 09 11:58:44 crc kubenswrapper[4770]: I1209 11:58:44.995424 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a853b3079ae327a5e2751c66ecf028c4a7e55fd1f1c6894af589afedfaf48e0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.001170 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanaaaf-account-delete-x5lgz"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.002323 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.002414 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02 is running failed: container process not found" containerID="819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.003505 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02 is running failed: container process not found" containerID="819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.003538 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-xhnq4" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.005888 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanaaaf-account-delete-x5lgz" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.008107 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010311 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010360 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010370 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement6c0f-account-delete-j5h7v" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010484 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010537 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.010588 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b744fb85c-jt5h9" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.012753 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.012855 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5674985874-bhmwx" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.020106 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data" (OuterVolumeSpecName: "config-data") pod "a9d1c6e5-c685-442c-94bb-a6f650adb837" (UID: "a9d1c6e5-c685-442c-94bb-a6f650adb837"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.032133 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.033492 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-z28k6"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.045259 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-z28k6"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.097067 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.097113 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.097123 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d1c6e5-c685-442c-94bb-a6f650adb837-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.100018 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2ce4-account-create-update-7d6qp"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.108680 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.115780 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2ce4-account-create-update-7d6qp"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.124822 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.164406 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-574ccd49f8-67m27"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.191328 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6x5x2"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.205331 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6x5x2"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.209172 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.220782 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.222931 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.226518 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.226766 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.226915 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.229459 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.536525 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.539179 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00b4541a-003a-4282-ad4e-4b1f24106f15" path="/var/lib/kubelet/pods/00b4541a-003a-4282-ad4e-4b1f24106f15/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.550771 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fc164b2-f1a7-4c38-be49-f53427b7fd91" path="/var/lib/kubelet/pods/0fc164b2-f1a7-4c38-be49-f53427b7fd91/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.552061 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" path="/var/lib/kubelet/pods/356f02ca-284f-4ad4-9709-a9b6fba60b88/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.552815 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c72d3ee-7e15-4402-940d-a4524531a11d" path="/var/lib/kubelet/pods/3c72d3ee-7e15-4402-940d-a4524531a11d/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.561793 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43a8a640-d610-4726-b2c6-84e99621d820" path="/var/lib/kubelet/pods/43a8a640-d610-4726-b2c6-84e99621d820/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.568244 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="518c58bd-4a2f-439f-b5f5-0eda21434884" path="/var/lib/kubelet/pods/518c58bd-4a2f-439f-b5f5-0eda21434884/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.569261 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.569325 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.569330 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f916059-5f99-4cb8-aac7-0f1c65f0eb1e" path="/var/lib/kubelet/pods/6f916059-5f99-4cb8-aac7-0f1c65f0eb1e/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.570366 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" path="/var/lib/kubelet/pods/846d1db4-8766-42c2-8fe4-3ed3ba6b80fb/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.570885 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="944b3781-77eb-4993-a42d-f483bb90ffea" path="/var/lib/kubelet/pods/944b3781-77eb-4993-a42d-f483bb90ffea/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.578487 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae" path="/var/lib/kubelet/pods/9b5c87b1-e0b3-4455-a2ad-a14e5bd241ae/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.587211 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a95d5818-27fc-440a-b086-54eefa13d3ab" path="/var/lib/kubelet/pods/a95d5818-27fc-440a-b086-54eefa13d3ab/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.592386 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad01631-8fd1-42c6-abcd-989631e88fd5" path="/var/lib/kubelet/pods/aad01631-8fd1-42c6-abcd-989631e88fd5/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.605443 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bffb1e84-2444-47b4-9152-1920771323c6" path="/var/lib/kubelet/pods/bffb1e84-2444-47b4-9152-1920771323c6/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.606457 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6363bd5-219b-4fb9-9695-7924791e3262" path="/var/lib/kubelet/pods/d6363bd5-219b-4fb9-9695-7924791e3262/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.613469 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d663e7bf-6267-4d24-a23b-ba2e6924bd0e" path="/var/lib/kubelet/pods/d663e7bf-6267-4d24-a23b-ba2e6924bd0e/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.616327 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2cc3bf-23a6-4b3f-a1fa-b845aad71787" path="/var/lib/kubelet/pods/de2cc3bf-23a6-4b3f-a1fa-b845aad71787/volumes" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.619396 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzpmn\" (UniqueName: \"kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn\") pod \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.619567 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts\") pod \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\" (UID: \"a7a5dd25-a6fa-4afa-b107-453682e98cbb\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.619608 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts\") pod \"fc74c6b9-5938-417d-bd71-062ecc932ced\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.620251 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29chb\" (UniqueName: \"kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb\") pod \"fc74c6b9-5938-417d-bd71-062ecc932ced\" (UID: \"fc74c6b9-5938-417d-bd71-062ecc932ced\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.621885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7a5dd25-a6fa-4afa-b107-453682e98cbb" (UID: "a7a5dd25-a6fa-4afa-b107-453682e98cbb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.621919 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc74c6b9-5938-417d-bd71-062ecc932ced" (UID: "fc74c6b9-5938-417d-bd71-062ecc932ced"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.629345 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn" (OuterVolumeSpecName: "kube-api-access-gzpmn") pod "a7a5dd25-a6fa-4afa-b107-453682e98cbb" (UID: "a7a5dd25-a6fa-4afa-b107-453682e98cbb"). InnerVolumeSpecName "kube-api-access-gzpmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.652289 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb" (OuterVolumeSpecName: "kube-api-access-29chb") pod "fc74c6b9-5938-417d-bd71-062ecc932ced" (UID: "fc74c6b9-5938-417d-bd71-062ecc932ced"). InnerVolumeSpecName "kube-api-access-29chb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.703590 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-078e-account-create-update-d6hrg"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.703641 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.721860 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29chb\" (UniqueName: \"kubernetes.io/projected/fc74c6b9-5938-417d-bd71-062ecc932ced-kube-api-access-29chb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.721912 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzpmn\" (UniqueName: \"kubernetes.io/projected/a7a5dd25-a6fa-4afa-b107-453682e98cbb-kube-api-access-gzpmn\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.721924 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a5dd25-a6fa-4afa-b107-453682e98cbb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.721937 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc74c6b9-5938-417d-bd71-062ecc932ced-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.747218 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:58:45 crc kubenswrapper[4770]: E1209 11:58:45.782589 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-xfqwf operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone9545-account-delete-8lfqw" podUID="9db9a008-f89f-44e5-9714-10deaf0fc649" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.818132 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-078e-account-create-update-d6hrg"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.827255 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.860727 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-44v45"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.927143 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs\") pod \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.927538 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts\") pod \"b1b663cc-0534-426b-ac89-cdd56dba00a5\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.927660 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-258z8\" (UniqueName: \"kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8\") pod \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.927746 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle\") pod \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.927833 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czxht\" (UniqueName: \"kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht\") pod \"b1b663cc-0534-426b-ac89-cdd56dba00a5\" (UID: \"b1b663cc-0534-426b-ac89-cdd56dba00a5\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.928003 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config\") pod \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\" (UID: \"732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b\") " Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.931916 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1b663cc-0534-426b-ac89-cdd56dba00a5" (UID: "b1b663cc-0534-426b-ac89-cdd56dba00a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.935107 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8" (OuterVolumeSpecName: "kube-api-access-258z8") pod "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" (UID: "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b"). InnerVolumeSpecName "kube-api-access-258z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.945634 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht" (OuterVolumeSpecName: "kube-api-access-czxht") pod "b1b663cc-0534-426b-ac89-cdd56dba00a5" (UID: "b1b663cc-0534-426b-ac89-cdd56dba00a5"). InnerVolumeSpecName "kube-api-access-czxht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.946055 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.979002 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-44v45"] Dec 09 11:58:45 crc kubenswrapper[4770]: I1209 11:58:45.992544 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.020934 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-17c1-account-create-update-vnk7f"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.024081 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6 is running failed: container process not found" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.033105 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6 is running failed: container process not found" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.034595 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" (UID: "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.034567 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6 is running failed: container process not found" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.034707 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.036231 4770 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.036264 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b663cc-0534-426b-ac89-cdd56dba00a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.036277 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-258z8\" (UniqueName: \"kubernetes.io/projected/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-api-access-258z8\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.036290 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czxht\" (UniqueName: \"kubernetes.io/projected/b1b663cc-0534-426b-ac89-cdd56dba00a5-kube-api-access-czxht\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.041581 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-17c1-account-create-update-vnk7f"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.058048 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" (UID: "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.060757 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-l8qd2"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.061948 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerID="bd1a0426233ccd1b1386234cfd700d23db549b81974daa36103a0f808577669b" exitCode=0 Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.062008 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerDied","Data":"bd1a0426233ccd1b1386234cfd700d23db549b81974daa36103a0f808577669b"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.064301 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" (UID: "732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.069152 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xhnq4_f754e609-fd35-4c45-bfe8-71c659a22cdb/ovn-controller/0.log" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.069212 4770 generic.go:334] "Generic (PLEG): container finished" podID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerID="819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" exitCode=137 Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.069287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4" event={"ID":"f754e609-fd35-4c45-bfe8-71c659a22cdb","Type":"ContainerDied","Data":"819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.069320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhnq4" event={"ID":"f754e609-fd35-4c45-bfe8-71c659a22cdb","Type":"ContainerDied","Data":"a1f58f7451278f7de20c41adf37fd92ef7d6e8e3761163d4d91c67e6e0ff8b1d"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.069333 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1f58f7451278f7de20c41adf37fd92ef7d6e8e3761163d4d91c67e6e0ff8b1d" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.072233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" event={"ID":"9f7b544a-d135-44fc-9fda-3938369f661f","Type":"ContainerDied","Data":"5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.072269 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b7332102027d6a1f03d03ba4aba17e045e422cc4fbc29f13b83f89701d6bde9" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.074192 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66bfc44b69-qfbmc" event={"ID":"568ef04e-032c-40f4-97b9-b21483281a4f","Type":"ContainerDied","Data":"0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.074224 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0839c7847434ce2335f40ab26d27e78fe6df21c3aa149a4e98d375d710c60c7f" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.079265 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-l8qd2"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.082270 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2ce4-account-delete-z5lmd" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.083220 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"df40ca35-a10e-4571-b460-f7cd465070d8","Type":"ContainerDied","Data":"b0f18994898aaa4cc5d7d372aa4ba2389d90e5d2cae9bac4e5de4333b9e7c985"} Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.083250 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0f18994898aaa4cc5d7d372aa4ba2389d90e5d2cae9bac4e5de4333b9e7c985" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.083301 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.084211 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.089138 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell017c1-account-delete-xwp6r" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.090566 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8c24-account-delete-ccctw" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.091061 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance078e-account-delete-lwfnm" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.091146 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.091120 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.091075 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder4e4b-account-delete-5tttq" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137268 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137316 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137376 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137427 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137472 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137551 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137592 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137662 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137706 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137774 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpnmk\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.137810 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf\") pod \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\" (UID: \"b596f863-b8d8-4fb4-93e3-45bdce78f1b1\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.138302 4770 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.138322 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.140404 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.139812 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.147042 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi8c24-account-delete-ccctw"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.149249 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.154962 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8c24-account-create-update-wxc89"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.158360 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.161979 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.162210 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info" (OuterVolumeSpecName: "pod-info") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.163895 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8c24-account-create-update-wxc89"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.174011 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi8c24-account-delete-ccctw"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.177781 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.183509 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.189101 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk" (OuterVolumeSpecName: "kube-api-access-jpnmk") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "kube-api-access-jpnmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.194390 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.202036 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.208302 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf" (OuterVolumeSpecName: "server-conf") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.210988 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6b744fb85c-jt5h9"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.242350 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246215 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246253 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246268 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246281 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246295 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpnmk\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-kube-api-access-jpnmk\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246310 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246321 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246332 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.246374 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.249946 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.254858 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data" (OuterVolumeSpecName: "config-data") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.257179 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.265941 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.266088 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.163:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.268263 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.272621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b596f863-b8d8-4fb4-93e3-45bdce78f1b1" (UID: "b596f863-b8d8-4fb4-93e3-45bdce78f1b1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.276805 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.283972 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.289841 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.304636 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.304847 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.317557 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.325397 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.336313 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.353717 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.354019 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b596f863-b8d8-4fb4-93e3-45bdce78f1b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.354068 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.357648 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.360674 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xhnq4_f754e609-fd35-4c45-bfe8-71c659a22cdb/ovn-controller/0.log" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.360740 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.381132 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.394870 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.404526 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.415746 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.422479 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5674985874-bhmwx"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454686 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454762 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data\") pod \"9f7b544a-d135-44fc-9fda-3938369f661f\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454816 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454857 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gggtc\" (UniqueName: \"kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc\") pod \"568ef04e-032c-40f4-97b9-b21483281a4f\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454891 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454941 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454960 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs\") pod \"9f7b544a-d135-44fc-9fda-3938369f661f\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.454987 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455008 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle\") pod \"568ef04e-032c-40f4-97b9-b21483281a4f\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455035 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455073 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vz5x\" (UniqueName: \"kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x\") pod \"9f7b544a-d135-44fc-9fda-3938369f661f\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn62r\" (UniqueName: \"kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r\") pod \"f352a11c-1887-4736-98cb-4eccf3086e97\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455123 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxvnd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455145 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data\") pod \"568ef04e-032c-40f4-97b9-b21483281a4f\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455208 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxkcb\" (UniqueName: \"kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb\") pod \"df40ca35-a10e-4571-b460-f7cd465070d8\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455227 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs\") pod \"568ef04e-032c-40f4-97b9-b21483281a4f\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455254 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455274 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data\") pod \"df40ca35-a10e-4571-b460-f7cd465070d8\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455315 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455358 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455384 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom\") pod \"9f7b544a-d135-44fc-9fda-3938369f661f\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455423 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle\") pod \"9f7b544a-d135-44fc-9fda-3938369f661f\" (UID: \"9f7b544a-d135-44fc-9fda-3938369f661f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455452 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data\") pod \"f352a11c-1887-4736-98cb-4eccf3086e97\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455475 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle\") pod \"df40ca35-a10e-4571-b460-f7cd465070d8\" (UID: \"df40ca35-a10e-4571-b460-f7cd465070d8\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455517 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom\") pod \"568ef04e-032c-40f4-97b9-b21483281a4f\" (UID: \"568ef04e-032c-40f4-97b9-b21483281a4f\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455558 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins\") pod \"03fef9de-3f78-48b3-9079-9dc87184f803\" (UID: \"03fef9de-3f78-48b3-9079-9dc87184f803\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455598 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle\") pod \"f352a11c-1887-4736-98cb-4eccf3086e97\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455625 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config\") pod \"f352a11c-1887-4736-98cb-4eccf3086e97\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.455668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs\") pod \"f352a11c-1887-4736-98cb-4eccf3086e97\" (UID: \"f352a11c-1887-4736-98cb-4eccf3086e97\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.456158 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.456281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.461886 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs" (OuterVolumeSpecName: "logs") pod "9f7b544a-d135-44fc-9fda-3938369f661f" (UID: "9f7b544a-d135-44fc-9fda-3938369f661f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.462876 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs" (OuterVolumeSpecName: "logs") pod "568ef04e-032c-40f4-97b9-b21483281a4f" (UID: "568ef04e-032c-40f4-97b9-b21483281a4f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.463707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.465826 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.466746 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f352a11c-1887-4736-98cb-4eccf3086e97" (UID: "f352a11c-1887-4736-98cb-4eccf3086e97"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.467283 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data" (OuterVolumeSpecName: "config-data") pod "f352a11c-1887-4736-98cb-4eccf3086e97" (UID: "f352a11c-1887-4736-98cb-4eccf3086e97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.469799 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.470171 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb" (OuterVolumeSpecName: "kube-api-access-fxkcb") pod "df40ca35-a10e-4571-b460-f7cd465070d8" (UID: "df40ca35-a10e-4571-b460-f7cd465070d8"). InnerVolumeSpecName "kube-api-access-fxkcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.473287 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd" (OuterVolumeSpecName: "kube-api-access-wxvnd") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "kube-api-access-wxvnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.480729 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.483541 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.484992 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x" (OuterVolumeSpecName: "kube-api-access-9vz5x") pod "9f7b544a-d135-44fc-9fda-3938369f661f" (UID: "9f7b544a-d135-44fc-9fda-3938369f661f"). InnerVolumeSpecName "kube-api-access-9vz5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.487198 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc" (OuterVolumeSpecName: "kube-api-access-gggtc") pod "568ef04e-032c-40f4-97b9-b21483281a4f" (UID: "568ef04e-032c-40f4-97b9-b21483281a4f"). InnerVolumeSpecName "kube-api-access-gggtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.491628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info" (OuterVolumeSpecName: "pod-info") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.496437 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9f7b544a-d135-44fc-9fda-3938369f661f" (UID: "9f7b544a-d135-44fc-9fda-3938369f661f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.501040 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.503210 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "568ef04e-032c-40f4-97b9-b21483281a4f" (UID: "568ef04e-032c-40f4-97b9-b21483281a4f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.519623 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.520117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r" (OuterVolumeSpecName: "kube-api-access-mn62r") pod "f352a11c-1887-4736-98cb-4eccf3086e97" (UID: "f352a11c-1887-4736-98cb-4eccf3086e97"). InnerVolumeSpecName "kube-api-access-mn62r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.534936 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.554340 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.563361 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.578366 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.578950 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579033 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579078 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579118 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9xkt\" (UniqueName: \"kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579147 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579182 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579231 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579307 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579346 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579374 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579431 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd\") pod \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\" (UID: \"f7e5fb15-de40-44d2-8e1c-fcb78a080d38\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579655 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj5lm\" (UniqueName: \"kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm\") pod \"f754e609-fd35-4c45-bfe8-71c659a22cdb\" (UID: \"f754e609-fd35-4c45-bfe8-71c659a22cdb\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579659 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts" (OuterVolumeSpecName: "scripts") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.579976 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run" (OuterVolumeSpecName: "var-run") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.580262 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.587644 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.594392 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data" (OuterVolumeSpecName: "config-data") pod "df40ca35-a10e-4571-b460-f7cd465070d8" (UID: "df40ca35-a10e-4571-b460-f7cd465070d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595139 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595167 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595178 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595189 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595221 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595232 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595242 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gggtc\" (UniqueName: \"kubernetes.io/projected/568ef04e-032c-40f4-97b9-b21483281a4f-kube-api-access-gggtc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595251 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03fef9de-3f78-48b3-9079-9dc87184f803-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595235 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm" (OuterVolumeSpecName: "kube-api-access-bj5lm") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "kube-api-access-bj5lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595260 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595332 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f7b544a-d135-44fc-9fda-3938369f661f-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595353 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03fef9de-3f78-48b3-9079-9dc87184f803-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595392 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595408 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vz5x\" (UniqueName: \"kubernetes.io/projected/9f7b544a-d135-44fc-9fda-3938369f661f-kube-api-access-9vz5x\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595424 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn62r\" (UniqueName: \"kubernetes.io/projected/f352a11c-1887-4736-98cb-4eccf3086e97-kube-api-access-mn62r\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595437 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxvnd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-kube-api-access-wxvnd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595452 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxkcb\" (UniqueName: \"kubernetes.io/projected/df40ca35-a10e-4571-b460-f7cd465070d8-kube-api-access-fxkcb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595467 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/568ef04e-032c-40f4-97b9-b21483281a4f-logs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595481 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595495 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595510 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f754e609-fd35-4c45-bfe8-71c659a22cdb-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595523 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595535 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f352a11c-1887-4736-98cb-4eccf3086e97-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595549 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.595611 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.596804 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.597185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f7b544a-d135-44fc-9fda-3938369f661f" (UID: "9f7b544a-d135-44fc-9fda-3938369f661f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.617501 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.622064 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts" (OuterVolumeSpecName: "scripts") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.623704 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt" (OuterVolumeSpecName: "kube-api-access-z9xkt") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "kube-api-access-z9xkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.668762 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df40ca35-a10e-4571-b460-f7cd465070d8" (UID: "df40ca35-a10e-4571-b460-f7cd465070d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.671564 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance078e-account-delete-lwfnm"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.685291 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "568ef04e-032c-40f4-97b9-b21483281a4f" (UID: "568ef04e-032c-40f4-97b9-b21483281a4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.697600 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.697663 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.697652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.697916 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.697964 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.698550 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.715293 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.716713 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzwjx\" (UniqueName: \"kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.716791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.716861 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.716964 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys\") pod \"b306554d-ccb5-4db1-8462-fef6aba58f57\" (UID: \"b306554d-ccb5-4db1-8462-fef6aba58f57\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.717699 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.717812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") pod \"keystone9545-account-delete-8lfqw\" (UID: \"9db9a008-f89f-44e5-9714-10deaf0fc649\") " pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718104 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718121 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718140 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df40ca35-a10e-4571-b460-f7cd465070d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718149 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9xkt\" (UniqueName: \"kubernetes.io/projected/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-kube-api-access-z9xkt\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718162 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718172 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718184 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f754e609-fd35-4c45-bfe8-71c659a22cdb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718194 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.718203 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj5lm\" (UniqueName: \"kubernetes.io/projected/f754e609-fd35-4c45-bfe8-71c659a22cdb-kube-api-access-bj5lm\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.718281 4770 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.718348 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:50.718332066 +0000 UTC m=+1655.959090585 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : configmap "openstack-scripts" not found Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.721424 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.726162 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.726244 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.728496 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.734526 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data" (OuterVolumeSpecName: "config-data") pod "568ef04e-032c-40f4-97b9-b21483281a4f" (UID: "568ef04e-032c-40f4-97b9-b21483281a4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.737314 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.739794 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx" (OuterVolumeSpecName: "kube-api-access-fzwjx") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "kube-api-access-fzwjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.743238 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.761880 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts" (OuterVolumeSpecName: "scripts") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.762114 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.763061 4770 projected.go:194] Error preparing data for projected volume kube-api-access-xfqwf for pod openstack/keystone9545-account-delete-8lfqw: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:46 crc kubenswrapper[4770]: E1209 11:58:46.764826 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf podName:9db9a008-f89f-44e5-9714-10deaf0fc649 nodeName:}" failed. No retries permitted until 2025-12-09 11:58:50.763295777 +0000 UTC m=+1656.004054296 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-xfqwf" (UniqueName: "kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf") pod "keystone9545-account-delete-8lfqw" (UID: "9db9a008-f89f-44e5-9714-10deaf0fc649") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.774742 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron2ce4-account-delete-z5lmd"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.792578 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data" (OuterVolumeSpecName: "config-data") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.796071 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.812650 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder4e4b-account-delete-5tttq"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.815418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "f352a11c-1887-4736-98cb-4eccf3086e97" (UID: "f352a11c-1887-4736-98cb-4eccf3086e97"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.818961 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle\") pod \"4e77fad7-c892-4d17-87ca-6c699f1b2258\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819077 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data\") pod \"4e77fad7-c892-4d17-87ca-6c699f1b2258\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819145 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmcg6\" (UniqueName: \"kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6\") pod \"4e77fad7-c892-4d17-87ca-6c699f1b2258\" (UID: \"4e77fad7-c892-4d17-87ca-6c699f1b2258\") " Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819488 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/568ef04e-032c-40f4-97b9-b21483281a4f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819508 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819518 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819526 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzwjx\" (UniqueName: \"kubernetes.io/projected/b306554d-ccb5-4db1-8462-fef6aba58f57-kube-api-access-fzwjx\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819539 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819547 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819555 4770 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.819566 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.824880 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6" (OuterVolumeSpecName: "kube-api-access-gmcg6") pod "4e77fad7-c892-4d17-87ca-6c699f1b2258" (UID: "4e77fad7-c892-4d17-87ca-6c699f1b2258"). InnerVolumeSpecName "kube-api-access-gmcg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.825654 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.827891 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.834915 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell017c1-account-delete-xwp6r"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.841064 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.851621 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.858655 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.858959 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.892118 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data" (OuterVolumeSpecName: "config-data") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.905240 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data" (OuterVolumeSpecName: "config-data") pod "9f7b544a-d135-44fc-9fda-3938369f661f" (UID: "9f7b544a-d135-44fc-9fda-3938369f661f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922249 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922297 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922310 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7b544a-d135-44fc-9fda-3938369f661f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922322 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922334 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.922346 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmcg6\" (UniqueName: \"kubernetes.io/projected/4e77fad7-c892-4d17-87ca-6c699f1b2258-kube-api-access-gmcg6\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.924924 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f352a11c-1887-4736-98cb-4eccf3086e97" (UID: "f352a11c-1887-4736-98cb-4eccf3086e97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.927362 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data" (OuterVolumeSpecName: "config-data") pod "4e77fad7-c892-4d17-87ca-6c699f1b2258" (UID: "4e77fad7-c892-4d17-87ca-6c699f1b2258"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.930352 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.939015 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.941206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "f754e609-fd35-4c45-bfe8-71c659a22cdb" (UID: "f754e609-fd35-4c45-bfe8-71c659a22cdb"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.951782 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf" (OuterVolumeSpecName: "server-conf") pod "03fef9de-3f78-48b3-9079-9dc87184f803" (UID: "03fef9de-3f78-48b3-9079-9dc87184f803"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.955680 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.961646 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e77fad7-c892-4d17-87ca-6c699f1b2258" (UID: "4e77fad7-c892-4d17-87ca-6c699f1b2258"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.966428 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b306554d-ccb5-4db1-8462-fef6aba58f57" (UID: "b306554d-ccb5-4db1-8462-fef6aba58f57"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:46 crc kubenswrapper[4770]: I1209 11:58:46.989069 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data" (OuterVolumeSpecName: "config-data") pod "f7e5fb15-de40-44d2-8e1c-fcb78a080d38" (UID: "f7e5fb15-de40-44d2-8e1c-fcb78a080d38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024430 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03fef9de-3f78-48b3-9079-9dc87184f803-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024478 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024489 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024498 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03fef9de-3f78-48b3-9079-9dc87184f803-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024507 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e77fad7-c892-4d17-87ca-6c699f1b2258-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.024515 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f754e609-fd35-4c45-bfe8-71c659a22cdb-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.025020 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.025030 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b306554d-ccb5-4db1-8462-fef6aba58f57-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.025065 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f352a11c-1887-4736-98cb-4eccf3086e97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.025085 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e5fb15-de40-44d2-8e1c-fcb78a080d38-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.060798 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.103521 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" exitCode=0 Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.103595 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.103644 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerDied","Data":"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.103732 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d6e15893-6d15-4fa7-abc0-33697cc9b4cc","Type":"ContainerDied","Data":"269faed7cb2f222c44ee3f00b36632affafba0f9d4af49831d837ec797f8e210"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.103772 4770 scope.go:117] "RemoveContainer" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.111308 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7e5fb15-de40-44d2-8e1c-fcb78a080d38","Type":"ContainerDied","Data":"6950ccae8d1bffad99d338e0ae384434148fe8c6b289dce6029995ac77c6226a"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.111444 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.118916 4770 generic.go:334] "Generic (PLEG): container finished" podID="b306554d-ccb5-4db1-8462-fef6aba58f57" containerID="97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919" exitCode=0 Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.119008 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b4585bb4d-2j2md" event={"ID":"b306554d-ccb5-4db1-8462-fef6aba58f57","Type":"ContainerDied","Data":"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.119411 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b4585bb4d-2j2md" event={"ID":"b306554d-ccb5-4db1-8462-fef6aba58f57","Type":"ContainerDied","Data":"2f5d6fb1acb5dbd5aca8d99eb0093122d83a493146fd4dc86ff66d2986c069d2"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.119087 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b4585bb4d-2j2md" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123066 4770 generic.go:334] "Generic (PLEG): container finished" podID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" exitCode=0 Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123200 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66bfc44b69-qfbmc" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123194 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123240 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-764f86c974-ns5jl" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e77fad7-c892-4d17-87ca-6c699f1b2258","Type":"ContainerDied","Data":"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123285 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e77fad7-c892-4d17-87ca-6c699f1b2258","Type":"ContainerDied","Data":"d466c3774d7cf034d786a2d6a4364f0240ad291c945c2f9e6d0322daf6b02e04"} Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123607 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123776 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123784 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123847 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhnq4" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.123857 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone9545-account-delete-8lfqw" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125689 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125809 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcddb\" (UniqueName: \"kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125842 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.125999 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.126060 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.127395 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.128138 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.129237 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.129848 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.144837 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb" (OuterVolumeSpecName: "kube-api-access-hcddb") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "kube-api-access-hcddb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.148327 4770 scope.go:117] "RemoveContainer" containerID="ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.411262 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.414012 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:58:47 crc kubenswrapper[4770]: E1209 11:58:47.414379 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.414673 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.414763 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\" (UID: \"d6e15893-6d15-4fa7-abc0-33697cc9b4cc\") " Dec 09 11:58:47 crc kubenswrapper[4770]: W1209 11:58:47.414991 4770 mount_helper_common.go:34] Warning: mount cleanup skipped because path does not exist: /var/lib/kubelet/pods/d6e15893-6d15-4fa7-abc0-33697cc9b4cc/volumes/kubernetes.io~local-volume/local-storage02-crc Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415020 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415336 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415355 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415368 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcddb\" (UniqueName: \"kubernetes.io/projected/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kube-api-access-hcddb\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415380 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415390 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415412 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.415423 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.420340 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" path="/var/lib/kubelet/pods/1c0a1baa-962d-4dfc-891e-3563feff00bf/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.422844 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e46e9fb-3610-414e-acef-28323578d294" path="/var/lib/kubelet/pods/1e46e9fb-3610-414e-acef-28323578d294/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.442350 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.445090 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edf704f-d8b6-4205-a818-bbcae6131f58" path="/var/lib/kubelet/pods/1edf704f-d8b6-4205-a818-bbcae6131f58/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.446353 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234ed53e-6da0-4fb5-8990-c4af4f500af8" path="/var/lib/kubelet/pods/234ed53e-6da0-4fb5-8990-c4af4f500af8/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.447789 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24be0696-be7e-46b9-aa90-d83d8abbd793" path="/var/lib/kubelet/pods/24be0696-be7e-46b9-aa90-d83d8abbd793/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.448628 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c47befa-1ab8-4d48-b43f-806e2335066a" path="/var/lib/kubelet/pods/2c47befa-1ab8-4d48-b43f-806e2335066a/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.450106 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" path="/var/lib/kubelet/pods/3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.450873 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" path="/var/lib/kubelet/pods/732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.451106 4770 scope.go:117] "RemoveContainer" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.454977 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" path="/var/lib/kubelet/pods/9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.456028 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1672e55-70d6-4e2e-9900-c2744265763b" path="/var/lib/kubelet/pods/a1672e55-70d6-4e2e-9900-c2744265763b/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.456194 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "d6e15893-6d15-4fa7-abc0-33697cc9b4cc" (UID: "d6e15893-6d15-4fa7-abc0-33697cc9b4cc"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:47 crc kubenswrapper[4770]: E1209 11:58:47.456707 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b\": container with ID starting with 79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b not found: ID does not exist" containerID="79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.456744 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b"} err="failed to get container status \"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b\": rpc error: code = NotFound desc = could not find container \"79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b\": container with ID starting with 79ac3708488bb82802eb7d0a88ddf3df5517fdfe213fb00201fd601bbf09d62b not found: ID does not exist" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.456789 4770 scope.go:117] "RemoveContainer" containerID="ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45" Dec 09 11:58:47 crc kubenswrapper[4770]: E1209 11:58:47.457210 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45\": container with ID starting with ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45 not found: ID does not exist" containerID="ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.457240 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45"} err="failed to get container status \"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45\": rpc error: code = NotFound desc = could not find container \"ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45\": container with ID starting with ada7ab0f8a3b979fa71a98cf6649300b4539ad5c8557d3c3611e14895149cf45 not found: ID does not exist" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.457259 4770 scope.go:117] "RemoveContainer" containerID="d3cff712577cb5ca4995f819406108099e3cabf583e0b9fe83d7a120676f4548" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.474275 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9" path="/var/lib/kubelet/pods/a6a8233f-447c-490d-9a3c-3cf8fcdfd9c9/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.477557 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a5dd25-a6fa-4afa-b107-453682e98cbb" path="/var/lib/kubelet/pods/a7a5dd25-a6fa-4afa-b107-453682e98cbb/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.478551 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" path="/var/lib/kubelet/pods/a9d1c6e5-c685-442c-94bb-a6f650adb837/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.490931 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1b663cc-0534-426b-ac89-cdd56dba00a5" path="/var/lib/kubelet/pods/b1b663cc-0534-426b-ac89-cdd56dba00a5/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.492872 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" path="/var/lib/kubelet/pods/b596f863-b8d8-4fb4-93e3-45bdce78f1b1/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.493805 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" path="/var/lib/kubelet/pods/c98e9f23-cf17-402d-8473-f10ba0c63a1d/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.495298 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" path="/var/lib/kubelet/pods/cead551d-3523-499a-a1b4-ebce9c97e59e/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.496083 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" path="/var/lib/kubelet/pods/dc923d25-3beb-45cd-a504-0ebb6391a9f4/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.496979 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" path="/var/lib/kubelet/pods/eb1b43fc-8d67-462b-a735-fcd38d19f7a8/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.502005 4770 scope.go:117] "RemoveContainer" containerID="b85b05cd17651d9b3bd353aa522c2eeb740da829a60de5831bcd5ea98ce4117d" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.502630 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" path="/var/lib/kubelet/pods/f17f6cc7-50ef-45f1-9552-edc09f1e6ecb/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.503321 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc74c6b9-5938-417d-bd71-062ecc932ced" path="/var/lib/kubelet/pods/fc74c6b9-5938-417d-bd71-062ecc932ced/volumes" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.505329 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.517622 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.517666 4770 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e15893-6d15-4fa7-abc0-33697cc9b4cc-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.524458 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.540638 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.548838 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.568390 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone9545-account-delete-8lfqw"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.584874 4770 scope.go:117] "RemoveContainer" containerID="bd1a0426233ccd1b1386234cfd700d23db549b81974daa36103a0f808577669b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.595233 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone9545-account-delete-8lfqw"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.609541 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.618708 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-66bfc44b69-qfbmc"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.619167 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfqwf\" (UniqueName: \"kubernetes.io/projected/9db9a008-f89f-44e5-9714-10deaf0fc649-kube-api-access-xfqwf\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.619193 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9db9a008-f89f-44e5-9714-10deaf0fc649-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.629799 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.655060 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.668362 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.682336 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.684635 4770 scope.go:117] "RemoveContainer" containerID="07ea87393e7d08e9bbfe345d3786c9fd9565d8d1f4171600282eccd2c3cc50f2" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.694201 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.704501 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.759318 4770 scope.go:117] "RemoveContainer" containerID="97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.764487 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.764568 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-764f86c974-ns5jl"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.777538 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.782468 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xhnq4"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.805347 4770 scope.go:117] "RemoveContainer" containerID="97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.805562 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:58:47 crc kubenswrapper[4770]: E1209 11:58:47.806430 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919\": container with ID starting with 97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919 not found: ID does not exist" containerID="97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.806576 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919"} err="failed to get container status \"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919\": rpc error: code = NotFound desc = could not find container \"97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919\": container with ID starting with 97433d2aa379753b9aae341b70becc9b4a36d9741a5b7af3eac0ac8c7bdac919 not found: ID does not exist" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.806851 4770 scope.go:117] "RemoveContainer" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.814580 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5b4585bb4d-2j2md"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.835023 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.851432 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.864143 4770 scope.go:117] "RemoveContainer" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" Dec 09 11:58:47 crc kubenswrapper[4770]: E1209 11:58:47.864999 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b\": container with ID starting with 96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b not found: ID does not exist" containerID="96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b" Dec 09 11:58:47 crc kubenswrapper[4770]: I1209 11:58:47.865133 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b"} err="failed to get container status \"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b\": rpc error: code = NotFound desc = could not find container \"96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b\": container with ID starting with 96301c35e7b89eb8ba02d5ebb713a8ebde126a4ecc41eb98d6ad330b4641985b not found: ID does not exist" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.157755 4770 generic.go:334] "Generic (PLEG): container finished" podID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerID="fd8527bc2e276f42630b31d0e4c2d54c775ef3cdb370d0ca709fc11bfc05e843" exitCode=0 Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.157832 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerDied","Data":"fd8527bc2e276f42630b31d0e4c2d54c775ef3cdb370d0ca709fc11bfc05e843"} Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.157866 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dc5b689b7-qgl92" event={"ID":"6ab03763-40e0-4e54-bb2a-7ac9de8c5184","Type":"ContainerDied","Data":"b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d"} Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.157881 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3045f5cae920d31ba1761b662c5bc36a0ec2424a531cc224c796d459c5c929d" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.162547 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_0c7c3022-77d1-4055-8d10-6c7a474a9833/ovn-northd/0.log" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.162586 4770 generic.go:334] "Generic (PLEG): container finished" podID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" exitCode=139 Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.162635 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerDied","Data":"867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870"} Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.163201 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.232950 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233050 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233082 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h5gw\" (UniqueName: \"kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233127 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233216 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.233307 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config\") pod \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\" (UID: \"6ab03763-40e0-4e54-bb2a-7ac9de8c5184\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.247559 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.247615 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw" (OuterVolumeSpecName: "kube-api-access-6h5gw") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "kube-api-access-6h5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.286173 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.291876 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.294409 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.294957 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config" (OuterVolumeSpecName: "config") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.334545 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6ab03763-40e0-4e54-bb2a-7ac9de8c5184" (UID: "6ab03763-40e0-4e54-bb2a-7ac9de8c5184"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336444 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336515 4770 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336532 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h5gw\" (UniqueName: \"kubernetes.io/projected/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-kube-api-access-6h5gw\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336548 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336562 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336576 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.336587 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6ab03763-40e0-4e54-bb2a-7ac9de8c5184-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.338742 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_0c7c3022-77d1-4055-8d10-6c7a474a9833/ovn-northd/0.log" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.338820 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.437986 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhpbn\" (UniqueName: \"kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438138 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438210 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438274 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438302 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438360 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.438433 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs\") pod \"0c7c3022-77d1-4055-8d10-6c7a474a9833\" (UID: \"0c7c3022-77d1-4055-8d10-6c7a474a9833\") " Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.439160 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.439439 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts" (OuterVolumeSpecName: "scripts") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.439952 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config" (OuterVolumeSpecName: "config") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.441804 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn" (OuterVolumeSpecName: "kube-api-access-nhpbn") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "kube-api-access-nhpbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.460285 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.493609 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.495922 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "0c7c3022-77d1-4055-8d10-6c7a474a9833" (UID: "0c7c3022-77d1-4055-8d10-6c7a474a9833"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539684 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539954 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539967 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539976 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539986 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhpbn\" (UniqueName: \"kubernetes.io/projected/0c7c3022-77d1-4055-8d10-6c7a474a9833-kube-api-access-nhpbn\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.539994 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3022-77d1-4055-8d10-6c7a474a9833-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.540001 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0c7c3022-77d1-4055-8d10-6c7a474a9833-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.646814 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: i/o timeout" Dec 09 11:58:48 crc kubenswrapper[4770]: I1209 11:58:48.646862 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5674985874-bhmwx" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.177790 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_0c7c3022-77d1-4055-8d10-6c7a474a9833/ovn-northd/0.log" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.177919 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dc5b689b7-qgl92" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.186455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0c7c3022-77d1-4055-8d10-6c7a474a9833","Type":"ContainerDied","Data":"962fa3b338d0c548a486cb5aa72063b0086c5248ec02a5ce01349ac200afd6a4"} Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.186723 4770 scope.go:117] "RemoveContainer" containerID="60b3561b63896f5bc9807fe5e6de66a0dadd48252dfe9a1b9a2f39c48715e956" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.186790 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.220527 4770 scope.go:117] "RemoveContainer" containerID="867c8693d5a8d51d504430fe5338b5b21e04ac63213ea70ac7d924e9173d5870" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.235312 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.246528 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6dc5b689b7-qgl92"] Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.257927 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.262080 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.352337 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" path="/var/lib/kubelet/pods/03fef9de-3f78-48b3-9079-9dc87184f803/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.353306 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" path="/var/lib/kubelet/pods/0c7c3022-77d1-4055-8d10-6c7a474a9833/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.354077 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" path="/var/lib/kubelet/pods/4e77fad7-c892-4d17-87ca-6c699f1b2258/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.355560 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" path="/var/lib/kubelet/pods/568ef04e-032c-40f4-97b9-b21483281a4f/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.356300 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" path="/var/lib/kubelet/pods/6ab03763-40e0-4e54-bb2a-7ac9de8c5184/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.356922 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db9a008-f89f-44e5-9714-10deaf0fc649" path="/var/lib/kubelet/pods/9db9a008-f89f-44e5-9714-10deaf0fc649/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.358064 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" path="/var/lib/kubelet/pods/9f7b544a-d135-44fc-9fda-3938369f661f/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.358984 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b306554d-ccb5-4db1-8462-fef6aba58f57" path="/var/lib/kubelet/pods/b306554d-ccb5-4db1-8462-fef6aba58f57/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.359848 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" path="/var/lib/kubelet/pods/d6e15893-6d15-4fa7-abc0-33697cc9b4cc/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.361367 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" path="/var/lib/kubelet/pods/df40ca35-a10e-4571-b460-f7cd465070d8/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.362105 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" path="/var/lib/kubelet/pods/f352a11c-1887-4736-98cb-4eccf3086e97/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.362804 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" path="/var/lib/kubelet/pods/f754e609-fd35-4c45-bfe8-71c659a22cdb/volumes" Dec 09 11:58:49 crc kubenswrapper[4770]: I1209 11:58:49.364312 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" path="/var/lib/kubelet/pods/f7e5fb15-de40-44d2-8e1c-fcb78a080d38/volumes" Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.221873 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.222531 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.222963 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.223037 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.223062 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.225612 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.231872 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:50 crc kubenswrapper[4770]: E1209 11:58:50.232202 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:58:51 crc kubenswrapper[4770]: I1209 11:58:51.151143 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.106:11211: i/o timeout" Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.219437 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.220342 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.220727 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.220826 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.221218 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.225457 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.229122 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:58:55 crc kubenswrapper[4770]: E1209 11:58:55.229240 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:58:59 crc kubenswrapper[4770]: I1209 11:58:59.341370 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:58:59 crc kubenswrapper[4770]: E1209 11:58:59.342145 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.219313 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.220449 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.221065 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.221127 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.221847 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.223422 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.224855 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:00 crc kubenswrapper[4770]: E1209 11:59:00.224914 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.219480 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.220388 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.220584 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.220752 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.220780 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.222426 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.224027 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 09 11:59:05 crc kubenswrapper[4770]: E1209 11:59:05.224074 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-gct95" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.363414 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.485750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache\") pod \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.485814 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock\") pod \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.485854 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.485946 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") pod \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.486029 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxcsl\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl\") pod \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\" (UID: \"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f\") " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.486988 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock" (OuterVolumeSpecName: "lock") pod "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.487585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache" (OuterVolumeSpecName: "cache") pod "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.491585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl" (OuterVolumeSpecName: "kube-api-access-nxcsl") pod "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f"). InnerVolumeSpecName "kube-api-access-nxcsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.491759 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.492022 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "swift") pod "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" (UID: "a0b34aa9-ae84-4b79-ba88-25c86cf5c51f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.518792 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gct95_5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2/ovs-vswitchd/0.log" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.519711 4770 generic.go:334] "Generic (PLEG): container finished" podID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" exitCode=137 Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.519765 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerDied","Data":"5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5"} Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.527993 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerID="e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6" exitCode=137 Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.528037 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6"} Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.528079 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a0b34aa9-ae84-4b79-ba88-25c86cf5c51f","Type":"ContainerDied","Data":"b10aa3a14105600acae388ed793d07042cdff68f5e1a6fb0546ab50d2924e520"} Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.528102 4770 scope.go:117] "RemoveContainer" containerID="e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.528121 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.591271 4770 scope.go:117] "RemoveContainer" containerID="13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.592192 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.592228 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxcsl\" (UniqueName: \"kubernetes.io/projected/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-kube-api-access-nxcsl\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.592238 4770 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-cache\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.592248 4770 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f-lock\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.592278 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.594340 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.600716 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.608526 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.613240 4770 scope.go:117] "RemoveContainer" containerID="80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.633801 4770 scope.go:117] "RemoveContainer" containerID="5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.653641 4770 scope.go:117] "RemoveContainer" containerID="3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.682208 4770 scope.go:117] "RemoveContainer" containerID="7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.694039 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.704711 4770 scope.go:117] "RemoveContainer" containerID="4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.724760 4770 scope.go:117] "RemoveContainer" containerID="4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.752454 4770 scope.go:117] "RemoveContainer" containerID="4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.780664 4770 scope.go:117] "RemoveContainer" containerID="9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.807182 4770 scope.go:117] "RemoveContainer" containerID="14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.837034 4770 scope.go:117] "RemoveContainer" containerID="d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.860138 4770 scope.go:117] "RemoveContainer" containerID="ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.883609 4770 scope.go:117] "RemoveContainer" containerID="87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.909670 4770 scope.go:117] "RemoveContainer" containerID="d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.933924 4770 scope.go:117] "RemoveContainer" containerID="e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.934870 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6\": container with ID starting with e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6 not found: ID does not exist" containerID="e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.934926 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6"} err="failed to get container status \"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6\": rpc error: code = NotFound desc = could not find container \"e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6\": container with ID starting with e29dd77bbd8c0d15eba2de0c9427fb61e2fdd82ea91e8877104fa9fc1d9ad6e6 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.934961 4770 scope.go:117] "RemoveContainer" containerID="13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.935233 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d\": container with ID starting with 13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d not found: ID does not exist" containerID="13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.935257 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d"} err="failed to get container status \"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d\": rpc error: code = NotFound desc = could not find container \"13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d\": container with ID starting with 13aefbd7e235d0c97403bd5fbe8bdbcd7faa25e357786b850e11786028e0376d not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.935274 4770 scope.go:117] "RemoveContainer" containerID="80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.935547 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0\": container with ID starting with 80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0 not found: ID does not exist" containerID="80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.935570 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0"} err="failed to get container status \"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0\": rpc error: code = NotFound desc = could not find container \"80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0\": container with ID starting with 80073e250f7e77bfd79bdc49a0f1c5f36d7da4fd06c41095bf149252f3e7b5b0 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.935589 4770 scope.go:117] "RemoveContainer" containerID="5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.936027 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0\": container with ID starting with 5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0 not found: ID does not exist" containerID="5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936061 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0"} err="failed to get container status \"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0\": rpc error: code = NotFound desc = could not find container \"5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0\": container with ID starting with 5d522cd25faee2e2b8a55911be62593877ecdfa069f23f1a0ca560327d588fe0 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936079 4770 scope.go:117] "RemoveContainer" containerID="3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.936525 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6\": container with ID starting with 3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6 not found: ID does not exist" containerID="3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936561 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6"} err="failed to get container status \"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6\": rpc error: code = NotFound desc = could not find container \"3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6\": container with ID starting with 3c25e2f73da63ca7296adac5e529124e2d7570e020d45b0f30a38cd311a543a6 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936584 4770 scope.go:117] "RemoveContainer" containerID="7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.936826 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d\": container with ID starting with 7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d not found: ID does not exist" containerID="7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936859 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d"} err="failed to get container status \"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d\": rpc error: code = NotFound desc = could not find container \"7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d\": container with ID starting with 7f85ffe4d9bdcaeac9127b25da73158f1189ab27dea8b9b73d9b5860a255068d not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.936878 4770 scope.go:117] "RemoveContainer" containerID="4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.937161 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7\": container with ID starting with 4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7 not found: ID does not exist" containerID="4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937191 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7"} err="failed to get container status \"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7\": rpc error: code = NotFound desc = could not find container \"4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7\": container with ID starting with 4be8dede523ae525a6ea9df3ee43350090d8bbe5bda7f1413506b91ffda347f7 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937210 4770 scope.go:117] "RemoveContainer" containerID="4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.937475 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1\": container with ID starting with 4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1 not found: ID does not exist" containerID="4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937508 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1"} err="failed to get container status \"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1\": rpc error: code = NotFound desc = could not find container \"4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1\": container with ID starting with 4d45d1f69f4f580dbfa9c25addc940d97ba63901144265751a6a3cee4f1903a1 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937529 4770 scope.go:117] "RemoveContainer" containerID="4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.937770 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab\": container with ID starting with 4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab not found: ID does not exist" containerID="4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937798 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab"} err="failed to get container status \"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab\": rpc error: code = NotFound desc = could not find container \"4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab\": container with ID starting with 4c846a54dd53eb750e49e39539a7af2effca800c25fb531c54921281e32723ab not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.937817 4770 scope.go:117] "RemoveContainer" containerID="9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.938112 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621\": container with ID starting with 9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621 not found: ID does not exist" containerID="9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938146 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621"} err="failed to get container status \"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621\": rpc error: code = NotFound desc = could not find container \"9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621\": container with ID starting with 9965ce17c87e8c921dbb510a162cbf23b391780d866cd58d1b7f7a7d8fbec621 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938164 4770 scope.go:117] "RemoveContainer" containerID="14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.938459 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364\": container with ID starting with 14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364 not found: ID does not exist" containerID="14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938492 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364"} err="failed to get container status \"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364\": rpc error: code = NotFound desc = could not find container \"14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364\": container with ID starting with 14ff33a867a79eba0dca6632f49b0d8a763282ec9904f4ba78afb8ab04708364 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938509 4770 scope.go:117] "RemoveContainer" containerID="d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.938776 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7\": container with ID starting with d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7 not found: ID does not exist" containerID="d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938805 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7"} err="failed to get container status \"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7\": rpc error: code = NotFound desc = could not find container \"d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7\": container with ID starting with d2c912e8b0310410abc92cdd3cc62834c7e7424a4fe91aba6e31f68e17e082c7 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.938823 4770 scope.go:117] "RemoveContainer" containerID="ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.939080 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713\": container with ID starting with ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713 not found: ID does not exist" containerID="ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.939111 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713"} err="failed to get container status \"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713\": rpc error: code = NotFound desc = could not find container \"ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713\": container with ID starting with ca87a1c28e2f6ed7583ed6bac8c5290945a562e57ba9f866a330e499591a9713 not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.939129 4770 scope.go:117] "RemoveContainer" containerID="87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.939401 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d\": container with ID starting with 87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d not found: ID does not exist" containerID="87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.939429 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d"} err="failed to get container status \"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d\": rpc error: code = NotFound desc = could not find container \"87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d\": container with ID starting with 87f242fbc36bdcefa77b8f423f0590083319498102cfa267292846a8beb43d5d not found: ID does not exist" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.939447 4770 scope.go:117] "RemoveContainer" containerID="d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3" Dec 09 11:59:06 crc kubenswrapper[4770]: E1209 11:59:06.939705 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3\": container with ID starting with d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3 not found: ID does not exist" containerID="d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3" Dec 09 11:59:06 crc kubenswrapper[4770]: I1209 11:59:06.939734 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3"} err="failed to get container status \"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3\": rpc error: code = NotFound desc = could not find container \"d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3\": container with ID starting with d31d2fba08486958b4af04b9c3b7b7198042289773ca70dd068732b7ee25a2a3 not found: ID does not exist" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.183171 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gct95_5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2/ovs-vswitchd/0.log" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.184829 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.302976 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303030 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303079 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4575p\" (UniqueName: \"kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303161 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log\") pod \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\" (UID: \"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2\") " Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303557 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303607 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log" (OuterVolumeSpecName: "var-log") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303627 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib" (OuterVolumeSpecName: "var-lib") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.303647 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run" (OuterVolumeSpecName: "var-run") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.304892 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts" (OuterVolumeSpecName: "scripts") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.307526 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p" (OuterVolumeSpecName: "kube-api-access-4575p") pod "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" (UID: "5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2"). InnerVolumeSpecName "kube-api-access-4575p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.354216 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" path="/var/lib/kubelet/pods/a0b34aa9-ae84-4b79-ba88-25c86cf5c51f/volumes" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405170 4770 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-log\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405203 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405214 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405222 4770 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405267 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4575p\" (UniqueName: \"kubernetes.io/projected/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-kube-api-access-4575p\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.405277 4770 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2-var-lib\") on node \"crc\" DevicePath \"\"" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.544424 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gct95_5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2/ovs-vswitchd/0.log" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.545727 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gct95" event={"ID":"5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2","Type":"ContainerDied","Data":"4cbc671a83e48d5acdf23fe813dbec0fda6b957ad84cbf364cb5f7e1cda1889f"} Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.545789 4770 scope.go:117] "RemoveContainer" containerID="5742fd9753e6cabdefde46fe103a0e612ae5daf766a04ba14f08c6e9072caac5" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.545827 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gct95" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.569592 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.572099 4770 scope.go:117] "RemoveContainer" containerID="37c8b1c05416c72921d6528889e737b274c186448e866f1a7767d05fdf819a3b" Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.575317 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-gct95"] Dec 09 11:59:07 crc kubenswrapper[4770]: I1209 11:59:07.594490 4770 scope.go:117] "RemoveContainer" containerID="ce5f3ea540f088f455b54f1fb0bf39f592ea7d27d976af881d2c46b7334b8cb7" Dec 09 11:59:09 crc kubenswrapper[4770]: I1209 11:59:09.352004 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" path="/var/lib/kubelet/pods/5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2/volumes" Dec 09 11:59:13 crc kubenswrapper[4770]: I1209 11:59:13.097444 4770 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf6188401-2bb4-4ddc-a097-6b99f99df9e8"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf6188401-2bb4-4ddc-a097-6b99f99df9e8] : Timed out while waiting for systemd to remove kubepods-besteffort-podf6188401_2bb4_4ddc_a097_6b99f99df9e8.slice" Dec 09 11:59:13 crc kubenswrapper[4770]: I1209 11:59:13.309445 4770 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod518c58bd-4a2f-439f-b5f5-0eda21434884"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod518c58bd-4a2f-439f-b5f5-0eda21434884] : Timed out while waiting for systemd to remove kubepods-besteffort-pod518c58bd_4a2f_439f_b5f5_0eda21434884.slice" Dec 09 11:59:13 crc kubenswrapper[4770]: I1209 11:59:13.340629 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:59:13 crc kubenswrapper[4770]: E1209 11:59:13.341027 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:59:14 crc kubenswrapper[4770]: I1209 11:59:14.262926 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.177:9292/healthcheck\": context deadline exceeded" Dec 09 11:59:14 crc kubenswrapper[4770]: I1209 11:59:14.262885 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.177:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.270363 4770 scope.go:117] "RemoveContainer" containerID="7d29da55d749c371a137286ed3944f43b67a190a65d6103836dab022c75fbc05" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.304924 4770 scope.go:117] "RemoveContainer" containerID="377a052b321d83d9e8bff8487f5444db77436defd96fd805994e7fdf82aecba9" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.331429 4770 scope.go:117] "RemoveContainer" containerID="bf9a900264b568b10dc3739abfa123854754954e74d507968ba4ce9a0b201378" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.340598 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:59:27 crc kubenswrapper[4770]: E1209 11:59:27.341067 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.369460 4770 scope.go:117] "RemoveContainer" containerID="2946bef94b26e9eb5227b621b0591908bb8f5db0d79016f7e9097a651629b7fc" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.402108 4770 scope.go:117] "RemoveContainer" containerID="9256f77b59bd0b11cf351f783c242ba7e3eb289025a5f44fd213395ee4bf575a" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.426371 4770 scope.go:117] "RemoveContainer" containerID="aaa58da97bffc14fa7574c9cd57935763dac5dc795f51d4a296070911eb74b14" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.460216 4770 scope.go:117] "RemoveContainer" containerID="d370b0cd44dd3020056530066f02dab0b0bb51657bf84b49247c082765bd2e1d" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.491520 4770 scope.go:117] "RemoveContainer" containerID="a7dc5c4c948b7e59408c0e3cf9bcfbf9cd1717522c72afefd9d893df384cec6a" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.516719 4770 scope.go:117] "RemoveContainer" containerID="74e47b78fc3704d9f254adeaf8bbb3db0c5dd90dcb22750ff79aaa07926d8e44" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.542703 4770 scope.go:117] "RemoveContainer" containerID="819791f5e357ba9924b9dccd3055343436065ed834d1e5771d25983a6fc15a02" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.566689 4770 scope.go:117] "RemoveContainer" containerID="ce7b985dd5d6ad69d34a28139df573f6e591655833d1d03998c27319b1dd79e8" Dec 09 11:59:27 crc kubenswrapper[4770]: I1209 11:59:27.588716 4770 scope.go:117] "RemoveContainer" containerID="010a70aaa353dcb251bebda226955ab3b511ce82dd40a3e21356735a99fc680e" Dec 09 11:59:42 crc kubenswrapper[4770]: I1209 11:59:42.340282 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:59:42 crc kubenswrapper[4770]: E1209 11:59:42.341090 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 11:59:53 crc kubenswrapper[4770]: I1209 11:59:53.340980 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 11:59:53 crc kubenswrapper[4770]: E1209 11:59:53.341808 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.153717 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm"] Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154668 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154687 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154703 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154710 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154726 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8a640-d610-4726-b2c6-84e99621d820" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154736 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8a640-d610-4726-b2c6-84e99621d820" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154755 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154763 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154778 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154786 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-server" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154797 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154804 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154816 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="setup-container" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154824 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="setup-container" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154834 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154842 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154857 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154866 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154881 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154889 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154923 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154931 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154944 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154952 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154960 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154967 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.154982 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24be0696-be7e-46b9-aa90-d83d8abbd793" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.154990 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="24be0696-be7e-46b9-aa90-d83d8abbd793" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155002 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-reaper" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155011 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-reaper" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155021 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155028 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155039 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155049 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155059 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234ed53e-6da0-4fb5-8990-c4af4f500af8" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155067 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="234ed53e-6da0-4fb5-8990-c4af4f500af8" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155077 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155084 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155096 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155103 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155113 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155119 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155131 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerName="nova-scheduler-scheduler" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155158 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerName="nova-scheduler-scheduler" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155171 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155179 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-server" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155193 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" containerName="memcached" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155201 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" containerName="memcached" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155211 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155218 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-server" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155225 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155233 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155256 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155264 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155275 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155283 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155292 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155299 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155314 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155321 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155332 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155340 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155354 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-expirer" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155361 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-expirer" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155373 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155382 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155398 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155405 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155414 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155421 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155429 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155436 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155445 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155467 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155485 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server-init" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155497 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server-init" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155509 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="rsync" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155522 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="rsync" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155530 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-central-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155540 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-central-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155549 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155556 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155565 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerName="kube-state-metrics" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155572 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerName="kube-state-metrics" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155585 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="openstack-network-exporter" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155594 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="openstack-network-exporter" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155602 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155610 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155622 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155629 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155640 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-notification-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155648 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-notification-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155660 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a5dd25-a6fa-4afa-b107-453682e98cbb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155668 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a5dd25-a6fa-4afa-b107-453682e98cbb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155676 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155684 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155695 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155702 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155711 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155717 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155730 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="mysql-bootstrap" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155738 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="mysql-bootstrap" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155747 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155754 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155761 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155768 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155778 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155785 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-server" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155793 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155800 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155809 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="sg-core" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155816 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="sg-core" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155829 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155836 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155852 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155860 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155869 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b663cc-0534-426b-ac89-cdd56dba00a5" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155877 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b663cc-0534-426b-ac89-cdd56dba00a5" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155885 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="swift-recon-cron" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155893 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="swift-recon-cron" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155924 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155933 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155946 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.155954 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.155964 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156125 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker-log" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.156167 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="mysql-bootstrap" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156178 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="mysql-bootstrap" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.156216 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="setup-container" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156224 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="setup-container" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.156242 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b306554d-ccb5-4db1-8462-fef6aba58f57" containerName="keystone-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156257 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b306554d-ccb5-4db1-8462-fef6aba58f57" containerName="keystone-api" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.156274 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc74c6b9-5938-417d-bd71-062ecc932ced" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156282 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc74c6b9-5938-417d-bd71-062ecc932ced" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: E1209 12:00:00.156294 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156303 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156736 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f754e609-fd35-4c45-bfe8-71c659a22cdb" containerName="ovn-controller" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156756 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156778 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156793 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e15893-6d15-4fa7-abc0-33697cc9b4cc" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156803 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovsdb-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156815 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8a640-d610-4726-b2c6-84e99621d820" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156825 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156837 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c0a1baa-962d-4dfc-891e-3563feff00bf" containerName="cinder-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156847 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156855 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="03fef9de-3f78-48b3-9079-9dc87184f803" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156865 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156875 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156959 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="24be0696-be7e-46b9-aa90-d83d8abbd793" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156975 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="ovn-northd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.156995 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157013 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-central-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157028 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-reaper" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157041 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7c3022-77d1-4055-8d10-6c7a474a9833" containerName="openstack-network-exporter" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157050 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157062 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e77fad7-c892-4d17-87ca-6c699f1b2258" containerName="nova-cell0-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157075 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157086 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bcd6bd6-1790-4c69-a2b4-7f4f50051b6b" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157100 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b596f863-b8d8-4fb4-93e3-45bdce78f1b1" containerName="rabbitmq" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157109 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="swift-recon-cron" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157122 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="732165e6-e9e6-4a7b-a3fb-d5f38c0b3a8b" containerName="kube-state-metrics" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157131 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="sg-core" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157144 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157157 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157167 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b663cc-0534-426b-ac89-cdd56dba00a5" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157178 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="rsync" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157192 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5969e8b4-92d6-4ef0-8a28-3a3b8ea10be2" containerName="ovs-vswitchd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157201 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157210 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cead551d-3523-499a-a1b4-ebce9c97e59e" containerName="nova-scheduler-scheduler" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157219 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c98e9f23-cf17-402d-8473-f10ba0c63a1d" containerName="nova-metadata-metadata" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157230 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc923d25-3beb-45cd-a504-0ebb6391a9f4" containerName="nova-api-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157240 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="356f02ca-284f-4ad4-9709-a9b6fba60b88" containerName="placement-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157248 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157263 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="234ed53e-6da0-4fb5-8990-c4af4f500af8" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157278 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157291 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-server" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157299 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157308 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a5dd25-a6fa-4afa-b107-453682e98cbb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157321 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-expirer" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157335 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-updater" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157348 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157357 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d1c6e5-c685-442c-94bb-a6f650adb837" containerName="glance-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157371 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-replicator" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157387 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e5fb15-de40-44d2-8e1c-fcb78a080d38" containerName="ceilometer-notification-agent" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157397 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="object-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157406 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17f6cc7-50ef-45f1-9552-edc09f1e6ecb" containerName="galera" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157416 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="account-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157426 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157438 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b43fc-8d67-462b-a735-fcd38d19f7a8" containerName="proxy-httpd" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157449 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc74c6b9-5938-417d-bd71-062ecc932ced" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157458 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f7b544a-d135-44fc-9fda-3938369f661f" containerName="barbican-keystone-listener" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157469 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="568ef04e-032c-40f4-97b9-b21483281a4f" containerName="barbican-worker" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157480 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b34aa9-ae84-4b79-ba88-25c86cf5c51f" containerName="container-auditor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157492 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c49f2a7-e6aa-4bb0-8c62-8aeda6f9aef3" containerName="barbican-api-log" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157509 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="df40ca35-a10e-4571-b460-f7cd465070d8" containerName="nova-cell1-conductor-conductor" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157522 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="846d1db4-8766-42c2-8fe4-3ed3ba6b80fb" containerName="mariadb-account-delete" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157532 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b306554d-ccb5-4db1-8462-fef6aba58f57" containerName="keystone-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157544 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab03763-40e0-4e54-bb2a-7ac9de8c5184" containerName="neutron-api" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.157552 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f352a11c-1887-4736-98cb-4eccf3086e97" containerName="memcached" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.158238 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.165148 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.165680 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.173333 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm"] Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.324468 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.324540 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6h2s\" (UniqueName: \"kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.324585 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.427557 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.427680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6h2s\" (UniqueName: \"kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.427760 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.429845 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.435684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.452550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6h2s\" (UniqueName: \"kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s\") pod \"collect-profiles-29421360-m2gjm\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:00 crc kubenswrapper[4770]: I1209 12:00:00.642085 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:01 crc kubenswrapper[4770]: I1209 12:00:01.148721 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm"] Dec 09 12:00:02 crc kubenswrapper[4770]: I1209 12:00:02.124353 4770 generic.go:334] "Generic (PLEG): container finished" podID="df1e7e38-0196-43f1-a5f3-b149ef929584" containerID="b7393b5b593402a6cad01634e574580bc6022d654143620f6f64057329e39af6" exitCode=0 Dec 09 12:00:02 crc kubenswrapper[4770]: I1209 12:00:02.124570 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" event={"ID":"df1e7e38-0196-43f1-a5f3-b149ef929584","Type":"ContainerDied","Data":"b7393b5b593402a6cad01634e574580bc6022d654143620f6f64057329e39af6"} Dec 09 12:00:02 crc kubenswrapper[4770]: I1209 12:00:02.124870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" event={"ID":"df1e7e38-0196-43f1-a5f3-b149ef929584","Type":"ContainerStarted","Data":"627504597d3377554e9218f5e820381dcd63ecbedf64a8864693e8445f922ae5"} Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.507370 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.630583 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6h2s\" (UniqueName: \"kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s\") pod \"df1e7e38-0196-43f1-a5f3-b149ef929584\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.631000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume\") pod \"df1e7e38-0196-43f1-a5f3-b149ef929584\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.631093 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume\") pod \"df1e7e38-0196-43f1-a5f3-b149ef929584\" (UID: \"df1e7e38-0196-43f1-a5f3-b149ef929584\") " Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.631697 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume" (OuterVolumeSpecName: "config-volume") pod "df1e7e38-0196-43f1-a5f3-b149ef929584" (UID: "df1e7e38-0196-43f1-a5f3-b149ef929584"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.635714 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df1e7e38-0196-43f1-a5f3-b149ef929584" (UID: "df1e7e38-0196-43f1-a5f3-b149ef929584"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.636078 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s" (OuterVolumeSpecName: "kube-api-access-j6h2s") pod "df1e7e38-0196-43f1-a5f3-b149ef929584" (UID: "df1e7e38-0196-43f1-a5f3-b149ef929584"). InnerVolumeSpecName "kube-api-access-j6h2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.732722 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df1e7e38-0196-43f1-a5f3-b149ef929584-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.732781 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df1e7e38-0196-43f1-a5f3-b149ef929584-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:00:03 crc kubenswrapper[4770]: I1209 12:00:03.732798 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6h2s\" (UniqueName: \"kubernetes.io/projected/df1e7e38-0196-43f1-a5f3-b149ef929584-kube-api-access-j6h2s\") on node \"crc\" DevicePath \"\"" Dec 09 12:00:04 crc kubenswrapper[4770]: I1209 12:00:04.143887 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" event={"ID":"df1e7e38-0196-43f1-a5f3-b149ef929584","Type":"ContainerDied","Data":"627504597d3377554e9218f5e820381dcd63ecbedf64a8864693e8445f922ae5"} Dec 09 12:00:04 crc kubenswrapper[4770]: I1209 12:00:04.144265 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="627504597d3377554e9218f5e820381dcd63ecbedf64a8864693e8445f922ae5" Dec 09 12:00:04 crc kubenswrapper[4770]: I1209 12:00:04.144024 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm" Dec 09 12:00:08 crc kubenswrapper[4770]: I1209 12:00:08.340528 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:00:08 crc kubenswrapper[4770]: E1209 12:00:08.341147 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:00:22 crc kubenswrapper[4770]: I1209 12:00:22.340641 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:00:22 crc kubenswrapper[4770]: E1209 12:00:22.341453 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.025620 4770 scope.go:117] "RemoveContainer" containerID="9941dbcdc944c557b193fbf074adcf6381fa10a6d213fce09f09f89ac8040b58" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.060839 4770 scope.go:117] "RemoveContainer" containerID="b090a5d51438df67dcd20b0f78e7f2772fbbf4a6a66ab96fd227723a8cab51f7" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.082562 4770 scope.go:117] "RemoveContainer" containerID="f2d4dc50bf2b36a841745128ed7fc9d07db8df1ab56a2af6bc7a83543d7f633f" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.108355 4770 scope.go:117] "RemoveContainer" containerID="1980ac00c990360135915c858d0f125e136b46aa153c5a4d5cf4019dd84f4355" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.127470 4770 scope.go:117] "RemoveContainer" containerID="036cd9d85dac93e7bd1195800ae605a32e7659f3dc8866a12f69f81295bfb20c" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.166634 4770 scope.go:117] "RemoveContainer" containerID="bd7eefa592658dc5f5efab697a12a1e89949a1ef3faff590424391e7c9cbec12" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.208356 4770 scope.go:117] "RemoveContainer" containerID="4123eed8acdc05741e19a06c69e7357fb168f03ee4b24e968baf996a6cf73961" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.230136 4770 scope.go:117] "RemoveContainer" containerID="846a9a013e8d368afaaf333fa475f4a8d0eece651d21d6b459c099f7297229b4" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.271092 4770 scope.go:117] "RemoveContainer" containerID="b70ca7d01822bd66dce685f79e1d82c3e5fbe093736ea216ac67d51a955e093b" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.292796 4770 scope.go:117] "RemoveContainer" containerID="84ccb36f1c40cfd48e74cef605f7a217a151d292ac3cb99fbc76392a343e5ec9" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.311270 4770 scope.go:117] "RemoveContainer" containerID="7d1eb31fddc9b44fd9d13bf2118bc46c56e247d2368c5da71f2f49e4c1c3bb1c" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.330175 4770 scope.go:117] "RemoveContainer" containerID="8d873c38c1b605a0973fce198ea3596a148416b4b267624233cff3d2a70e4864" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.384578 4770 scope.go:117] "RemoveContainer" containerID="c22d19cca7b2713787bd65a3a2f28d2818e0752a37bcb6ec50e1587fb6477014" Dec 09 12:00:28 crc kubenswrapper[4770]: I1209 12:00:28.409551 4770 scope.go:117] "RemoveContainer" containerID="cd1933388921d9c04dede556a2c178263df5ffaf0f51165517bcc4205f9bbbac" Dec 09 12:00:35 crc kubenswrapper[4770]: I1209 12:00:35.352271 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:00:35 crc kubenswrapper[4770]: E1209 12:00:35.353064 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:00:50 crc kubenswrapper[4770]: I1209 12:00:50.340388 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:00:50 crc kubenswrapper[4770]: E1209 12:00:50.342398 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:01:03 crc kubenswrapper[4770]: I1209 12:01:03.342085 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:01:03 crc kubenswrapper[4770]: E1209 12:01:03.342819 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:01:18 crc kubenswrapper[4770]: I1209 12:01:18.343275 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:01:18 crc kubenswrapper[4770]: E1209 12:01:18.343966 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.667953 4770 scope.go:117] "RemoveContainer" containerID="fd8527bc2e276f42630b31d0e4c2d54c775ef3cdb370d0ca709fc11bfc05e843" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.695577 4770 scope.go:117] "RemoveContainer" containerID="297de169838c4bac5ceb5ffe3cb876b689e8b9332529573ff7910aa430b83e44" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.718015 4770 scope.go:117] "RemoveContainer" containerID="9b65050fb96dad2b68de130b2e74efdc4a249da8425436217ee12e2727d54d98" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.762162 4770 scope.go:117] "RemoveContainer" containerID="da61d729c0b6b1e02f1c0a6ae649a12798627a0b0d85b763d039b8fe64855854" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.786047 4770 scope.go:117] "RemoveContainer" containerID="28cd7242a9c9640679276453e171ec22b26b2720ffe571f2f85b629ee46a5e06" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.812313 4770 scope.go:117] "RemoveContainer" containerID="0ed885e5102a41a0b21a56c6ea93cb492355c4740fed766c91160632cf4c5c43" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.831308 4770 scope.go:117] "RemoveContainer" containerID="b505b07675385a501eefd16b47033f1d0cdb0b120345d67e47e86e33ef0f2548" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.849014 4770 scope.go:117] "RemoveContainer" containerID="4ba87b24c19c1120bbcabbace77c866035e530eaf47c87572eba2d7e47526454" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.867943 4770 scope.go:117] "RemoveContainer" containerID="6909006db23528ef5901641d836bab057983dc7c9e64e0113cbd39ca6d6fab1d" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.915523 4770 scope.go:117] "RemoveContainer" containerID="85925850bf61e5c8f64a81b9f9e5037d619b24f1d0912e5127b4a2874bb652ee" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.947458 4770 scope.go:117] "RemoveContainer" containerID="60819c01211cdf868787f2091ef5192fb54f532e2c708f8cbbacdddd49d63cda" Dec 09 12:01:28 crc kubenswrapper[4770]: I1209 12:01:28.996755 4770 scope.go:117] "RemoveContainer" containerID="b398d9c6572a1c1d9989fa5b3cdc1de7aed57f3c833bda8fd8efb536cd6ce5db" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.025634 4770 scope.go:117] "RemoveContainer" containerID="12e1f31a90f3e82dc874c093f9ae7defb528a89431960cfdaa426ccc760094c3" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.045468 4770 scope.go:117] "RemoveContainer" containerID="948f5d8ee8bd59bbc4c4f416e35f1ff0d5497a183addd81d6baefcacf2368a39" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.063567 4770 scope.go:117] "RemoveContainer" containerID="e77c216dc47b6c7c9a614efb4c3cf8153cbd683bc99a04574e90b93b09aa208a" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.080607 4770 scope.go:117] "RemoveContainer" containerID="c5ec8fa6eed14138a68a6b229d128065bd09d1ea7c60899215996437c5d9b880" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.098147 4770 scope.go:117] "RemoveContainer" containerID="74aa53e345d67730abe1cb4d00a4a413ff9c3de93027ad63aa017d74368659fe" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.125239 4770 scope.go:117] "RemoveContainer" containerID="bf8ad42774200d477849ea268cd29b0c65b5c006ef52af1b6710676014a7e842" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.144719 4770 scope.go:117] "RemoveContainer" containerID="7bc283a32d2b16f74f1cf489c744ab75e9f7ab17baf7e2b5ac315e76c1b74b42" Dec 09 12:01:29 crc kubenswrapper[4770]: I1209 12:01:29.162939 4770 scope.go:117] "RemoveContainer" containerID="3390dfb8cbb1f0c6bd1a75c816ce715ca36a3164d095376f2127e25ddb826b67" Dec 09 12:01:30 crc kubenswrapper[4770]: I1209 12:01:30.340851 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:01:30 crc kubenswrapper[4770]: E1209 12:01:30.342360 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:01:42 crc kubenswrapper[4770]: I1209 12:01:42.341162 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:01:42 crc kubenswrapper[4770]: E1209 12:01:42.342062 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:01:53 crc kubenswrapper[4770]: I1209 12:01:53.339955 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:01:53 crc kubenswrapper[4770]: E1209 12:01:53.340728 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:02:04 crc kubenswrapper[4770]: I1209 12:02:04.344681 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:02:04 crc kubenswrapper[4770]: E1209 12:02:04.345475 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:02:15 crc kubenswrapper[4770]: I1209 12:02:15.401585 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:02:15 crc kubenswrapper[4770]: E1209 12:02:15.402388 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.340368 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:02:29 crc kubenswrapper[4770]: E1209 12:02:29.341623 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.347295 4770 scope.go:117] "RemoveContainer" containerID="b2d9d48cbe19f172b7ac2c4f7e5d87d39111563cf778bcbc8050188702d7c75f" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.375995 4770 scope.go:117] "RemoveContainer" containerID="f483d04a4a2d6845118ac8581f5705bae9286fde5e96eb516d573219ca6e36da" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.426149 4770 scope.go:117] "RemoveContainer" containerID="b6dee3f17438b10b86b6871f88dfe7765fc45e68f4da390d513c0bd71565c8c6" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.447192 4770 scope.go:117] "RemoveContainer" containerID="b4612188bb4e931c0a88e007267426d57b3f35243474f53c7fa6fbe2b96d7e90" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.475619 4770 scope.go:117] "RemoveContainer" containerID="baec3dc745b1f2b4180b68e8ccb0a514d518f4ffcd36092950684e12e387a258" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.493585 4770 scope.go:117] "RemoveContainer" containerID="2bbe5d78041ffbdd917a78be6b066422f3130937304df0fb5318bb5b4e09ea2f" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.524602 4770 scope.go:117] "RemoveContainer" containerID="fe68553e529f03896be808d62dfc73b3485de5dc39de9017e810c7baf0d4521c" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.544779 4770 scope.go:117] "RemoveContainer" containerID="fa7bcf45aabcd3aa07a85a0fa2eb580f33813d2af8b453470cd5519a7b3285cd" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.635925 4770 scope.go:117] "RemoveContainer" containerID="07184d4ac9451f9079837acb457f303a341323b0ecb029c30cb51673a1719aae" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.657825 4770 scope.go:117] "RemoveContainer" containerID="8ce8e611661071a5e49fc120f1a9bef7f615630c825216f4eed7d8cb555e5d6c" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.679864 4770 scope.go:117] "RemoveContainer" containerID="0193e4873f3244cab191ab4be299640f973319dfcf035f612a78cdfcd54c17a7" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.700004 4770 scope.go:117] "RemoveContainer" containerID="eebed45a512ca372586490859e850a999b9d581d64647896a65dfb070e1af5fb" Dec 09 12:02:29 crc kubenswrapper[4770]: I1209 12:02:29.721658 4770 scope.go:117] "RemoveContainer" containerID="7569e5bc00921bcf0e4cb0006c71b05ba900b55b24a670f35b3c7271e62f80ca" Dec 09 12:02:40 crc kubenswrapper[4770]: I1209 12:02:40.340237 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:02:40 crc kubenswrapper[4770]: E1209 12:02:40.341034 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:02:54 crc kubenswrapper[4770]: I1209 12:02:54.340453 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:02:54 crc kubenswrapper[4770]: E1209 12:02:54.341257 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:03:08 crc kubenswrapper[4770]: I1209 12:03:08.340501 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:03:08 crc kubenswrapper[4770]: E1209 12:03:08.341243 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:03:23 crc kubenswrapper[4770]: I1209 12:03:23.340365 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:03:23 crc kubenswrapper[4770]: E1209 12:03:23.341147 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:03:29 crc kubenswrapper[4770]: I1209 12:03:29.869180 4770 scope.go:117] "RemoveContainer" containerID="eaf666a59707d09ab24d8771ad533795627f89bf69142f938756699b674ac7a2" Dec 09 12:03:29 crc kubenswrapper[4770]: I1209 12:03:29.898935 4770 scope.go:117] "RemoveContainer" containerID="218393e2da0bbfb5410ec4757c8cb027246a3f4ac476769670c88ceb23a30731" Dec 09 12:03:29 crc kubenswrapper[4770]: I1209 12:03:29.920668 4770 scope.go:117] "RemoveContainer" containerID="5c4f62f25d849a06be8caf3d78db3b07c172a14e1cf6ef3ea14bd2facea76ab6" Dec 09 12:03:29 crc kubenswrapper[4770]: I1209 12:03:29.943790 4770 scope.go:117] "RemoveContainer" containerID="b221dc48fbe0244448017160d5eb9b96fc984cc9361787e38ba36c7d6f5633ab" Dec 09 12:03:29 crc kubenswrapper[4770]: I1209 12:03:29.997146 4770 scope.go:117] "RemoveContainer" containerID="1292e06e7f5b284f4f90cb329ddfb73c6b447dce5d6ada412db691e074d9d6e8" Dec 09 12:03:30 crc kubenswrapper[4770]: I1209 12:03:30.044167 4770 scope.go:117] "RemoveContainer" containerID="ae7aead3b7c5a3e9931ebf74de8ed5ead9c4d5c4cf68dcd3f1e07a9eaa054b77" Dec 09 12:03:30 crc kubenswrapper[4770]: I1209 12:03:30.063129 4770 scope.go:117] "RemoveContainer" containerID="5116de47af635a5b00f7da13b387beee7aab34ff8ea4ca434e4f6c2e3090bf91" Dec 09 12:03:30 crc kubenswrapper[4770]: I1209 12:03:30.080000 4770 scope.go:117] "RemoveContainer" containerID="7da7aaafcf09e00c4474192a91af6510a6ae9de2e1122d1925a532e82d5a71a3" Dec 09 12:03:37 crc kubenswrapper[4770]: I1209 12:03:37.341377 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:03:38 crc kubenswrapper[4770]: I1209 12:03:38.241763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3"} Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.175212 4770 scope.go:117] "RemoveContainer" containerID="9a473cbcc51156e661f62a9cb155beb880dd96bd5e5e0ee00faf14c8c0beeb93" Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.205288 4770 scope.go:117] "RemoveContainer" containerID="603f5bb321d03410442e30b0c6f4a276f62b4480a2b95aa1ef1c48a81bca45e8" Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.227165 4770 scope.go:117] "RemoveContainer" containerID="b18e180734f74a73d57257fdf1cd1f9761bbb9036ddc64718a4744fc937a04d5" Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.249256 4770 scope.go:117] "RemoveContainer" containerID="243be21689d2a29c2966994727f43b6692216066e84a9a1e11711c1c2553ffb8" Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.267913 4770 scope.go:117] "RemoveContainer" containerID="8e69aba5ad2edf194e1fbfdc44ca9cb279f74fad72141a9489720b1480d0b44c" Dec 09 12:04:30 crc kubenswrapper[4770]: I1209 12:04:30.360839 4770 scope.go:117] "RemoveContainer" containerID="2acefd5c93d69b4a3fbb89f99aaa0a66050c911f86c19e2e8287fddec8fcebc8" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.443119 4770 scope.go:117] "RemoveContainer" containerID="f7d03dfb52ad2f0872647ee11eede8299ffd7bddfcea9b460d3aa7f5fd2f65b2" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.482454 4770 scope.go:117] "RemoveContainer" containerID="e4009481f228df01d9b43e603422c8d7f751e18820ed452cc044d226f068d5c1" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.500943 4770 scope.go:117] "RemoveContainer" containerID="b72abc12a6683d231d24f05fd066d588593a15880270b01d02054a0969cc24af" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.526176 4770 scope.go:117] "RemoveContainer" containerID="7aeca38e4fe31f32f1d0868743e6f73a07261e01e2d06fdcd6544df45818681d" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.553446 4770 scope.go:117] "RemoveContainer" containerID="245450190d772ae583ccda78ae92d3463c2252d79f297de71d0637f77ad6613f" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.575324 4770 scope.go:117] "RemoveContainer" containerID="66009135b64a083604820eb9655df507a1476d6aad063c1eff80d8350570d286" Dec 09 12:05:30 crc kubenswrapper[4770]: I1209 12:05:30.602842 4770 scope.go:117] "RemoveContainer" containerID="d42d0e251a4da94628ca8a8846c80d5cce60c71ad7836c652d111407395f41ec" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.579114 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:05:46 crc kubenswrapper[4770]: E1209 12:05:46.585323 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df1e7e38-0196-43f1-a5f3-b149ef929584" containerName="collect-profiles" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.585390 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="df1e7e38-0196-43f1-a5f3-b149ef929584" containerName="collect-profiles" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.585711 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="df1e7e38-0196-43f1-a5f3-b149ef929584" containerName="collect-profiles" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.587680 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.596706 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.708804 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.708858 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfqqw\" (UniqueName: \"kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.709027 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.810288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.810342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfqqw\" (UniqueName: \"kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.810402 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.811034 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.811035 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.831222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfqqw\" (UniqueName: \"kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw\") pod \"redhat-operators-hclsz\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:46 crc kubenswrapper[4770]: I1209 12:05:46.922557 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:47 crc kubenswrapper[4770]: I1209 12:05:47.398419 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:05:47 crc kubenswrapper[4770]: I1209 12:05:47.773945 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerStarted","Data":"f11c2a2744c05ca4141fda47e616c64417d6f71052486126dca694ff3c1d7865"} Dec 09 12:05:48 crc kubenswrapper[4770]: I1209 12:05:48.826086 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9556b46-a112-4b55-b242-a4f086327694" containerID="91efe4de8fd4e9d47a67dd1a715c109a6662435c345441c3b54f5713410b9125" exitCode=0 Dec 09 12:05:48 crc kubenswrapper[4770]: I1209 12:05:48.826150 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerDied","Data":"91efe4de8fd4e9d47a67dd1a715c109a6662435c345441c3b54f5713410b9125"} Dec 09 12:05:48 crc kubenswrapper[4770]: I1209 12:05:48.828313 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 12:05:49 crc kubenswrapper[4770]: I1209 12:05:49.834311 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerStarted","Data":"322de4b6dfd65d795fc942308f66fdc05c6435ce7517827bacbb7e14bd484f75"} Dec 09 12:05:50 crc kubenswrapper[4770]: I1209 12:05:50.848653 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9556b46-a112-4b55-b242-a4f086327694" containerID="322de4b6dfd65d795fc942308f66fdc05c6435ce7517827bacbb7e14bd484f75" exitCode=0 Dec 09 12:05:50 crc kubenswrapper[4770]: I1209 12:05:50.848710 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerDied","Data":"322de4b6dfd65d795fc942308f66fdc05c6435ce7517827bacbb7e14bd484f75"} Dec 09 12:05:51 crc kubenswrapper[4770]: I1209 12:05:51.866484 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerStarted","Data":"37b63ae59d0889d95d5b1724b1f481e5ae6ad4f84dba00b58063487a5ad21344"} Dec 09 12:05:51 crc kubenswrapper[4770]: I1209 12:05:51.891075 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hclsz" podStartSLOduration=3.468148174 podStartE2EDuration="5.891039263s" podCreationTimestamp="2025-12-09 12:05:46 +0000 UTC" firstStartedPulling="2025-12-09 12:05:48.827923325 +0000 UTC m=+2074.068681844" lastFinishedPulling="2025-12-09 12:05:51.250814414 +0000 UTC m=+2076.491572933" observedRunningTime="2025-12-09 12:05:51.885511904 +0000 UTC m=+2077.126270423" watchObservedRunningTime="2025-12-09 12:05:51.891039263 +0000 UTC m=+2077.131797792" Dec 09 12:05:56 crc kubenswrapper[4770]: I1209 12:05:56.922864 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:56 crc kubenswrapper[4770]: I1209 12:05:56.923842 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:56 crc kubenswrapper[4770]: I1209 12:05:56.973180 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:57 crc kubenswrapper[4770]: I1209 12:05:57.961249 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:05:58 crc kubenswrapper[4770]: I1209 12:05:58.018314 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:05:59 crc kubenswrapper[4770]: I1209 12:05:59.920114 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hclsz" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="registry-server" containerID="cri-o://37b63ae59d0889d95d5b1724b1f481e5ae6ad4f84dba00b58063487a5ad21344" gracePeriod=2 Dec 09 12:06:02 crc kubenswrapper[4770]: I1209 12:06:02.474215 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:06:02 crc kubenswrapper[4770]: I1209 12:06:02.475121 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.009709 4770 generic.go:334] "Generic (PLEG): container finished" podID="b9556b46-a112-4b55-b242-a4f086327694" containerID="37b63ae59d0889d95d5b1724b1f481e5ae6ad4f84dba00b58063487a5ad21344" exitCode=0 Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.009764 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerDied","Data":"37b63ae59d0889d95d5b1724b1f481e5ae6ad4f84dba00b58063487a5ad21344"} Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.166728 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.353690 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfqqw\" (UniqueName: \"kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw\") pod \"b9556b46-a112-4b55-b242-a4f086327694\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.353866 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content\") pod \"b9556b46-a112-4b55-b242-a4f086327694\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.353915 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities\") pod \"b9556b46-a112-4b55-b242-a4f086327694\" (UID: \"b9556b46-a112-4b55-b242-a4f086327694\") " Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.354744 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities" (OuterVolumeSpecName: "utilities") pod "b9556b46-a112-4b55-b242-a4f086327694" (UID: "b9556b46-a112-4b55-b242-a4f086327694"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.370203 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw" (OuterVolumeSpecName: "kube-api-access-cfqqw") pod "b9556b46-a112-4b55-b242-a4f086327694" (UID: "b9556b46-a112-4b55-b242-a4f086327694"). InnerVolumeSpecName "kube-api-access-cfqqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.456314 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.456351 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfqqw\" (UniqueName: \"kubernetes.io/projected/b9556b46-a112-4b55-b242-a4f086327694-kube-api-access-cfqqw\") on node \"crc\" DevicePath \"\"" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.530535 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9556b46-a112-4b55-b242-a4f086327694" (UID: "b9556b46-a112-4b55-b242-a4f086327694"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:06:04 crc kubenswrapper[4770]: I1209 12:06:04.557552 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9556b46-a112-4b55-b242-a4f086327694-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.019805 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hclsz" event={"ID":"b9556b46-a112-4b55-b242-a4f086327694","Type":"ContainerDied","Data":"f11c2a2744c05ca4141fda47e616c64417d6f71052486126dca694ff3c1d7865"} Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.019868 4770 scope.go:117] "RemoveContainer" containerID="37b63ae59d0889d95d5b1724b1f481e5ae6ad4f84dba00b58063487a5ad21344" Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.019918 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hclsz" Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.040155 4770 scope.go:117] "RemoveContainer" containerID="322de4b6dfd65d795fc942308f66fdc05c6435ce7517827bacbb7e14bd484f75" Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.060181 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.067977 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hclsz"] Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.071084 4770 scope.go:117] "RemoveContainer" containerID="91efe4de8fd4e9d47a67dd1a715c109a6662435c345441c3b54f5713410b9125" Dec 09 12:06:05 crc kubenswrapper[4770]: I1209 12:06:05.353432 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9556b46-a112-4b55-b242-a4f086327694" path="/var/lib/kubelet/pods/b9556b46-a112-4b55-b242-a4f086327694/volumes" Dec 09 12:06:32 crc kubenswrapper[4770]: I1209 12:06:32.473680 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:06:32 crc kubenswrapper[4770]: I1209 12:06:32.474334 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.475132 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.476011 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.476112 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.477002 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.477183 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3" gracePeriod=600 Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.880812 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3" exitCode=0 Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.881197 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3"} Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.881328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a"} Dec 09 12:07:02 crc kubenswrapper[4770]: I1209 12:07:02.881434 4770 scope.go:117] "RemoveContainer" containerID="319a867635228cf7f2cb4bd8dd2c8acfc056800355c8999db9feae6aa0152118" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.116834 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:25 crc kubenswrapper[4770]: E1209 12:07:25.118621 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="extract-content" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.118650 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="extract-content" Dec 09 12:07:25 crc kubenswrapper[4770]: E1209 12:07:25.118717 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="registry-server" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.118727 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="registry-server" Dec 09 12:07:25 crc kubenswrapper[4770]: E1209 12:07:25.118752 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="extract-utilities" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.118761 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="extract-utilities" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.119309 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9556b46-a112-4b55-b242-a4f086327694" containerName="registry-server" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.122848 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.129242 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.282961 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.283617 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.283743 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nzt9\" (UniqueName: \"kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.385220 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.385319 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nzt9\" (UniqueName: \"kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.385401 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.385844 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.385863 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.416916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nzt9\" (UniqueName: \"kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9\") pod \"redhat-marketplace-tv9jz\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.495864 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:25 crc kubenswrapper[4770]: I1209 12:07:25.993819 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:26 crc kubenswrapper[4770]: I1209 12:07:26.178123 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerStarted","Data":"2ede8a838f38f73a8cc1af878db5bc70821c194ed45663bf7f8b80891fa83beb"} Dec 09 12:07:27 crc kubenswrapper[4770]: I1209 12:07:27.187546 4770 generic.go:334] "Generic (PLEG): container finished" podID="f1810bc9-c225-4d64-906a-2c555a136a88" containerID="41d043918d2aceffe2bd8b649727bd6256b441d5f294fa0edd7f64e8de2fa9cb" exitCode=0 Dec 09 12:07:27 crc kubenswrapper[4770]: I1209 12:07:27.187590 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerDied","Data":"41d043918d2aceffe2bd8b649727bd6256b441d5f294fa0edd7f64e8de2fa9cb"} Dec 09 12:07:29 crc kubenswrapper[4770]: I1209 12:07:29.203670 4770 generic.go:334] "Generic (PLEG): container finished" podID="f1810bc9-c225-4d64-906a-2c555a136a88" containerID="6c675e2416e021899184c7d85f21b506e93ebc6db919bc411684ccbaa7b7e5f9" exitCode=0 Dec 09 12:07:29 crc kubenswrapper[4770]: I1209 12:07:29.203893 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerDied","Data":"6c675e2416e021899184c7d85f21b506e93ebc6db919bc411684ccbaa7b7e5f9"} Dec 09 12:07:32 crc kubenswrapper[4770]: I1209 12:07:32.229066 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerStarted","Data":"70116aa113e14b5ade2b3bafd883f16148c2bb1f0d61887d389b5c1fb605540e"} Dec 09 12:07:32 crc kubenswrapper[4770]: I1209 12:07:32.256364 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tv9jz" podStartSLOduration=3.7794932169999997 podStartE2EDuration="7.256330213s" podCreationTimestamp="2025-12-09 12:07:25 +0000 UTC" firstStartedPulling="2025-12-09 12:07:27.189940932 +0000 UTC m=+2172.430699451" lastFinishedPulling="2025-12-09 12:07:30.666777928 +0000 UTC m=+2175.907536447" observedRunningTime="2025-12-09 12:07:32.248326331 +0000 UTC m=+2177.489084850" watchObservedRunningTime="2025-12-09 12:07:32.256330213 +0000 UTC m=+2177.497088732" Dec 09 12:07:35 crc kubenswrapper[4770]: I1209 12:07:35.496802 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:35 crc kubenswrapper[4770]: I1209 12:07:35.497377 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:35 crc kubenswrapper[4770]: I1209 12:07:35.552046 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:36 crc kubenswrapper[4770]: I1209 12:07:36.309115 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:36 crc kubenswrapper[4770]: I1209 12:07:36.355868 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:38 crc kubenswrapper[4770]: I1209 12:07:38.281878 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tv9jz" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="registry-server" containerID="cri-o://70116aa113e14b5ade2b3bafd883f16148c2bb1f0d61887d389b5c1fb605540e" gracePeriod=2 Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.293876 4770 generic.go:334] "Generic (PLEG): container finished" podID="f1810bc9-c225-4d64-906a-2c555a136a88" containerID="70116aa113e14b5ade2b3bafd883f16148c2bb1f0d61887d389b5c1fb605540e" exitCode=0 Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.293949 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerDied","Data":"70116aa113e14b5ade2b3bafd883f16148c2bb1f0d61887d389b5c1fb605540e"} Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.364452 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.484162 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities\") pod \"f1810bc9-c225-4d64-906a-2c555a136a88\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.484349 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nzt9\" (UniqueName: \"kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9\") pod \"f1810bc9-c225-4d64-906a-2c555a136a88\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.484409 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content\") pod \"f1810bc9-c225-4d64-906a-2c555a136a88\" (UID: \"f1810bc9-c225-4d64-906a-2c555a136a88\") " Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.485327 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities" (OuterVolumeSpecName: "utilities") pod "f1810bc9-c225-4d64-906a-2c555a136a88" (UID: "f1810bc9-c225-4d64-906a-2c555a136a88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.490881 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9" (OuterVolumeSpecName: "kube-api-access-4nzt9") pod "f1810bc9-c225-4d64-906a-2c555a136a88" (UID: "f1810bc9-c225-4d64-906a-2c555a136a88"). InnerVolumeSpecName "kube-api-access-4nzt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.508158 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1810bc9-c225-4d64-906a-2c555a136a88" (UID: "f1810bc9-c225-4d64-906a-2c555a136a88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.585997 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nzt9\" (UniqueName: \"kubernetes.io/projected/f1810bc9-c225-4d64-906a-2c555a136a88-kube-api-access-4nzt9\") on node \"crc\" DevicePath \"\"" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.586039 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:07:39 crc kubenswrapper[4770]: I1209 12:07:39.586052 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1810bc9-c225-4d64-906a-2c555a136a88-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.306289 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tv9jz" event={"ID":"f1810bc9-c225-4d64-906a-2c555a136a88","Type":"ContainerDied","Data":"2ede8a838f38f73a8cc1af878db5bc70821c194ed45663bf7f8b80891fa83beb"} Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.306354 4770 scope.go:117] "RemoveContainer" containerID="70116aa113e14b5ade2b3bafd883f16148c2bb1f0d61887d389b5c1fb605540e" Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.306359 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tv9jz" Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.325428 4770 scope.go:117] "RemoveContainer" containerID="6c675e2416e021899184c7d85f21b506e93ebc6db919bc411684ccbaa7b7e5f9" Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.348074 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.353020 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tv9jz"] Dec 09 12:07:40 crc kubenswrapper[4770]: I1209 12:07:40.369229 4770 scope.go:117] "RemoveContainer" containerID="41d043918d2aceffe2bd8b649727bd6256b441d5f294fa0edd7f64e8de2fa9cb" Dec 09 12:07:41 crc kubenswrapper[4770]: I1209 12:07:41.348385 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" path="/var/lib/kubelet/pods/f1810bc9-c225-4d64-906a-2c555a136a88/volumes" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.212525 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:07:56 crc kubenswrapper[4770]: E1209 12:07:56.213244 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="registry-server" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.213263 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="registry-server" Dec 09 12:07:56 crc kubenswrapper[4770]: E1209 12:07:56.213283 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="extract-utilities" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.213294 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="extract-utilities" Dec 09 12:07:56 crc kubenswrapper[4770]: E1209 12:07:56.213327 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="extract-content" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.213337 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="extract-content" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.213528 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1810bc9-c225-4d64-906a-2c555a136a88" containerName="registry-server" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.214966 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.223616 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.558000 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnl4j\" (UniqueName: \"kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.558078 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.558126 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.659397 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnl4j\" (UniqueName: \"kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.659754 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.659795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.660357 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.661415 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.690004 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnl4j\" (UniqueName: \"kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j\") pod \"certified-operators-cbh8x\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:56 crc kubenswrapper[4770]: I1209 12:07:56.837670 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:07:58 crc kubenswrapper[4770]: I1209 12:07:58.014924 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:07:58 crc kubenswrapper[4770]: I1209 12:07:58.639049 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerID="b72df88d54d14af2d253a62c34910782e4eb13b2450ba10bfa1d6a5ab2efa15a" exitCode=0 Dec 09 12:07:58 crc kubenswrapper[4770]: I1209 12:07:58.639135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerDied","Data":"b72df88d54d14af2d253a62c34910782e4eb13b2450ba10bfa1d6a5ab2efa15a"} Dec 09 12:07:58 crc kubenswrapper[4770]: I1209 12:07:58.639429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerStarted","Data":"d36b3d43ed680de6beabd8d93c997e65831922b69cd1d00151b9eefe493357b2"} Dec 09 12:08:01 crc kubenswrapper[4770]: I1209 12:08:01.663760 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerStarted","Data":"b7ef5f0c7281113e51b6e45e66e24113c6a280f7d796337e79974709fcde824e"} Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.281234 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.283009 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.294083 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.386249 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5zj\" (UniqueName: \"kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.386348 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.386386 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.488201 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.488624 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.488724 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.488861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5zj\" (UniqueName: \"kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.489046 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.507953 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5zj\" (UniqueName: \"kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj\") pod \"community-operators-l9l59\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.628456 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.672604 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerID="b7ef5f0c7281113e51b6e45e66e24113c6a280f7d796337e79974709fcde824e" exitCode=0 Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.672651 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerDied","Data":"b7ef5f0c7281113e51b6e45e66e24113c6a280f7d796337e79974709fcde824e"} Dec 09 12:08:02 crc kubenswrapper[4770]: I1209 12:08:02.908257 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:02 crc kubenswrapper[4770]: W1209 12:08:02.912339 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62cfe03e_73b5_4639_a892_4659f5212936.slice/crio-8fc0198969f8a9443c1ea0da9c712e68de49e8e291c7365bc821fa39fab5c431 WatchSource:0}: Error finding container 8fc0198969f8a9443c1ea0da9c712e68de49e8e291c7365bc821fa39fab5c431: Status 404 returned error can't find the container with id 8fc0198969f8a9443c1ea0da9c712e68de49e8e291c7365bc821fa39fab5c431 Dec 09 12:08:03 crc kubenswrapper[4770]: I1209 12:08:03.682970 4770 generic.go:334] "Generic (PLEG): container finished" podID="62cfe03e-73b5-4639-a892-4659f5212936" containerID="27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7" exitCode=0 Dec 09 12:08:03 crc kubenswrapper[4770]: I1209 12:08:03.683029 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerDied","Data":"27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7"} Dec 09 12:08:03 crc kubenswrapper[4770]: I1209 12:08:03.683058 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerStarted","Data":"8fc0198969f8a9443c1ea0da9c712e68de49e8e291c7365bc821fa39fab5c431"} Dec 09 12:08:06 crc kubenswrapper[4770]: I1209 12:08:06.709840 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerStarted","Data":"39aad20f5845f3b06c69b41d353d92ce90df0c6a2be2ed0aa8f9fda66803ba35"} Dec 09 12:08:06 crc kubenswrapper[4770]: I1209 12:08:06.712996 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerStarted","Data":"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a"} Dec 09 12:08:06 crc kubenswrapper[4770]: I1209 12:08:06.741967 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cbh8x" podStartSLOduration=3.798092471 podStartE2EDuration="10.741949504s" podCreationTimestamp="2025-12-09 12:07:56 +0000 UTC" firstStartedPulling="2025-12-09 12:07:58.647605814 +0000 UTC m=+2203.888364333" lastFinishedPulling="2025-12-09 12:08:05.591462847 +0000 UTC m=+2210.832221366" observedRunningTime="2025-12-09 12:08:06.736795954 +0000 UTC m=+2211.977554473" watchObservedRunningTime="2025-12-09 12:08:06.741949504 +0000 UTC m=+2211.982708023" Dec 09 12:08:06 crc kubenswrapper[4770]: I1209 12:08:06.839131 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:06 crc kubenswrapper[4770]: I1209 12:08:06.839175 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:07 crc kubenswrapper[4770]: I1209 12:08:07.880466 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cbh8x" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="registry-server" probeResult="failure" output=< Dec 09 12:08:07 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 12:08:07 crc kubenswrapper[4770]: > Dec 09 12:08:09 crc kubenswrapper[4770]: I1209 12:08:09.737051 4770 generic.go:334] "Generic (PLEG): container finished" podID="62cfe03e-73b5-4639-a892-4659f5212936" containerID="9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a" exitCode=0 Dec 09 12:08:09 crc kubenswrapper[4770]: I1209 12:08:09.737148 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerDied","Data":"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a"} Dec 09 12:08:10 crc kubenswrapper[4770]: I1209 12:08:10.747119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerStarted","Data":"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c"} Dec 09 12:08:10 crc kubenswrapper[4770]: I1209 12:08:10.775807 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l9l59" podStartSLOduration=3.2541076 podStartE2EDuration="8.775787088s" podCreationTimestamp="2025-12-09 12:08:02 +0000 UTC" firstStartedPulling="2025-12-09 12:08:04.696486966 +0000 UTC m=+2209.937245495" lastFinishedPulling="2025-12-09 12:08:10.218166464 +0000 UTC m=+2215.458924983" observedRunningTime="2025-12-09 12:08:10.764597826 +0000 UTC m=+2216.005356345" watchObservedRunningTime="2025-12-09 12:08:10.775787088 +0000 UTC m=+2216.016545607" Dec 09 12:08:12 crc kubenswrapper[4770]: I1209 12:08:12.629116 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:12 crc kubenswrapper[4770]: I1209 12:08:12.630176 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:12 crc kubenswrapper[4770]: I1209 12:08:12.680097 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:16 crc kubenswrapper[4770]: I1209 12:08:16.886654 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:16 crc kubenswrapper[4770]: I1209 12:08:16.940156 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:17 crc kubenswrapper[4770]: I1209 12:08:17.132996 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:08:18 crc kubenswrapper[4770]: I1209 12:08:18.814234 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cbh8x" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="registry-server" containerID="cri-o://39aad20f5845f3b06c69b41d353d92ce90df0c6a2be2ed0aa8f9fda66803ba35" gracePeriod=2 Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.831139 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerID="39aad20f5845f3b06c69b41d353d92ce90df0c6a2be2ed0aa8f9fda66803ba35" exitCode=0 Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.831254 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerDied","Data":"39aad20f5845f3b06c69b41d353d92ce90df0c6a2be2ed0aa8f9fda66803ba35"} Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.920492 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.964822 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content\") pod \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.965028 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities\") pod \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.965105 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnl4j\" (UniqueName: \"kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j\") pod \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\" (UID: \"ae8d9807-4248-4897-a1e0-c62c6e27b38e\") " Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.974487 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities" (OuterVolumeSpecName: "utilities") pod "ae8d9807-4248-4897-a1e0-c62c6e27b38e" (UID: "ae8d9807-4248-4897-a1e0-c62c6e27b38e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.974873 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:19 crc kubenswrapper[4770]: I1209 12:08:19.987280 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j" (OuterVolumeSpecName: "kube-api-access-lnl4j") pod "ae8d9807-4248-4897-a1e0-c62c6e27b38e" (UID: "ae8d9807-4248-4897-a1e0-c62c6e27b38e"). InnerVolumeSpecName "kube-api-access-lnl4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.033422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae8d9807-4248-4897-a1e0-c62c6e27b38e" (UID: "ae8d9807-4248-4897-a1e0-c62c6e27b38e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.076317 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8d9807-4248-4897-a1e0-c62c6e27b38e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.076368 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnl4j\" (UniqueName: \"kubernetes.io/projected/ae8d9807-4248-4897-a1e0-c62c6e27b38e-kube-api-access-lnl4j\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.843087 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbh8x" event={"ID":"ae8d9807-4248-4897-a1e0-c62c6e27b38e","Type":"ContainerDied","Data":"d36b3d43ed680de6beabd8d93c997e65831922b69cd1d00151b9eefe493357b2"} Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.843167 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbh8x" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.843476 4770 scope.go:117] "RemoveContainer" containerID="39aad20f5845f3b06c69b41d353d92ce90df0c6a2be2ed0aa8f9fda66803ba35" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.870179 4770 scope.go:117] "RemoveContainer" containerID="b7ef5f0c7281113e51b6e45e66e24113c6a280f7d796337e79974709fcde824e" Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.878053 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.887129 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cbh8x"] Dec 09 12:08:20 crc kubenswrapper[4770]: I1209 12:08:20.901652 4770 scope.go:117] "RemoveContainer" containerID="b72df88d54d14af2d253a62c34910782e4eb13b2450ba10bfa1d6a5ab2efa15a" Dec 09 12:08:21 crc kubenswrapper[4770]: I1209 12:08:21.349764 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" path="/var/lib/kubelet/pods/ae8d9807-4248-4897-a1e0-c62c6e27b38e/volumes" Dec 09 12:08:22 crc kubenswrapper[4770]: I1209 12:08:22.687469 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:22 crc kubenswrapper[4770]: I1209 12:08:22.740128 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:22 crc kubenswrapper[4770]: I1209 12:08:22.859050 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l9l59" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="registry-server" containerID="cri-o://d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c" gracePeriod=2 Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.276097 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.333513 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb5zj\" (UniqueName: \"kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj\") pod \"62cfe03e-73b5-4639-a892-4659f5212936\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.333767 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content\") pod \"62cfe03e-73b5-4639-a892-4659f5212936\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.333831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities\") pod \"62cfe03e-73b5-4639-a892-4659f5212936\" (UID: \"62cfe03e-73b5-4639-a892-4659f5212936\") " Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.335185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities" (OuterVolumeSpecName: "utilities") pod "62cfe03e-73b5-4639-a892-4659f5212936" (UID: "62cfe03e-73b5-4639-a892-4659f5212936"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.340973 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj" (OuterVolumeSpecName: "kube-api-access-gb5zj") pod "62cfe03e-73b5-4639-a892-4659f5212936" (UID: "62cfe03e-73b5-4639-a892-4659f5212936"). InnerVolumeSpecName "kube-api-access-gb5zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.392781 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62cfe03e-73b5-4639-a892-4659f5212936" (UID: "62cfe03e-73b5-4639-a892-4659f5212936"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.436101 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.436190 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb5zj\" (UniqueName: \"kubernetes.io/projected/62cfe03e-73b5-4639-a892-4659f5212936-kube-api-access-gb5zj\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.436208 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cfe03e-73b5-4639-a892-4659f5212936-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.871754 4770 generic.go:334] "Generic (PLEG): container finished" podID="62cfe03e-73b5-4639-a892-4659f5212936" containerID="d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c" exitCode=0 Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.871808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerDied","Data":"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c"} Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.871944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l59" event={"ID":"62cfe03e-73b5-4639-a892-4659f5212936","Type":"ContainerDied","Data":"8fc0198969f8a9443c1ea0da9c712e68de49e8e291c7365bc821fa39fab5c431"} Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.871992 4770 scope.go:117] "RemoveContainer" containerID="d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.873234 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l59" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.893979 4770 scope.go:117] "RemoveContainer" containerID="9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.912166 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.920643 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l9l59"] Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.927739 4770 scope.go:117] "RemoveContainer" containerID="27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.949109 4770 scope.go:117] "RemoveContainer" containerID="d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c" Dec 09 12:08:23 crc kubenswrapper[4770]: E1209 12:08:23.949679 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c\": container with ID starting with d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c not found: ID does not exist" containerID="d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.949723 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c"} err="failed to get container status \"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c\": rpc error: code = NotFound desc = could not find container \"d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c\": container with ID starting with d09f664cd3ad495efd11d35f4a3c60e93e6e391b0b28c91608ab28489c3b632c not found: ID does not exist" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.949749 4770 scope.go:117] "RemoveContainer" containerID="9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a" Dec 09 12:08:23 crc kubenswrapper[4770]: E1209 12:08:23.950208 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a\": container with ID starting with 9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a not found: ID does not exist" containerID="9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.950254 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a"} err="failed to get container status \"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a\": rpc error: code = NotFound desc = could not find container \"9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a\": container with ID starting with 9e1037823f648b130b6858972df9e6e752d7054cf10fa96386229a728467418a not found: ID does not exist" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.950284 4770 scope.go:117] "RemoveContainer" containerID="27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7" Dec 09 12:08:23 crc kubenswrapper[4770]: E1209 12:08:23.950671 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7\": container with ID starting with 27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7 not found: ID does not exist" containerID="27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7" Dec 09 12:08:23 crc kubenswrapper[4770]: I1209 12:08:23.950698 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7"} err="failed to get container status \"27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7\": rpc error: code = NotFound desc = could not find container \"27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7\": container with ID starting with 27ca9c2a1bda1fa990db1e21471e30f1e22b57d3c25c21408bc7e75b61cdaab7 not found: ID does not exist" Dec 09 12:08:25 crc kubenswrapper[4770]: I1209 12:08:25.355203 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62cfe03e-73b5-4639-a892-4659f5212936" path="/var/lib/kubelet/pods/62cfe03e-73b5-4639-a892-4659f5212936/volumes" Dec 09 12:09:02 crc kubenswrapper[4770]: I1209 12:09:02.473750 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:09:02 crc kubenswrapper[4770]: I1209 12:09:02.474209 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:09:32 crc kubenswrapper[4770]: I1209 12:09:32.474236 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:09:32 crc kubenswrapper[4770]: I1209 12:09:32.474611 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.474216 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.474860 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.475002 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.476047 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.476141 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" gracePeriod=600 Dec 09 12:10:02 crc kubenswrapper[4770]: E1209 12:10:02.600175 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.721837 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" exitCode=0 Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.721941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a"} Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.722005 4770 scope.go:117] "RemoveContainer" containerID="27e09c88760fcca6b5e4099cff530e1bf6b9f6b95bfcbf2da8b65e1fb77328c3" Dec 09 12:10:02 crc kubenswrapper[4770]: I1209 12:10:02.722934 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:10:02 crc kubenswrapper[4770]: E1209 12:10:02.723427 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:10:13 crc kubenswrapper[4770]: I1209 12:10:13.341063 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:10:13 crc kubenswrapper[4770]: E1209 12:10:13.341963 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:10:25 crc kubenswrapper[4770]: I1209 12:10:25.345460 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:10:25 crc kubenswrapper[4770]: E1209 12:10:25.346321 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:10:36 crc kubenswrapper[4770]: I1209 12:10:36.340670 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:10:36 crc kubenswrapper[4770]: E1209 12:10:36.341532 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:10:48 crc kubenswrapper[4770]: I1209 12:10:48.340594 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:10:48 crc kubenswrapper[4770]: E1209 12:10:48.341340 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:11:03 crc kubenswrapper[4770]: I1209 12:11:03.340302 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:11:03 crc kubenswrapper[4770]: E1209 12:11:03.340888 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:11:17 crc kubenswrapper[4770]: I1209 12:11:17.340977 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:11:17 crc kubenswrapper[4770]: E1209 12:11:17.342937 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:11:31 crc kubenswrapper[4770]: I1209 12:11:31.340147 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:11:31 crc kubenswrapper[4770]: E1209 12:11:31.340888 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:11:44 crc kubenswrapper[4770]: I1209 12:11:44.340613 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:11:44 crc kubenswrapper[4770]: E1209 12:11:44.341412 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:11:56 crc kubenswrapper[4770]: I1209 12:11:56.340327 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:11:56 crc kubenswrapper[4770]: E1209 12:11:56.341000 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:12:08 crc kubenswrapper[4770]: I1209 12:12:08.340121 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:12:08 crc kubenswrapper[4770]: E1209 12:12:08.341160 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:12:19 crc kubenswrapper[4770]: I1209 12:12:19.340690 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:12:19 crc kubenswrapper[4770]: E1209 12:12:19.341441 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:12:30 crc kubenswrapper[4770]: I1209 12:12:30.340361 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:12:30 crc kubenswrapper[4770]: E1209 12:12:30.341610 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:12:42 crc kubenswrapper[4770]: I1209 12:12:42.340672 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:12:42 crc kubenswrapper[4770]: E1209 12:12:42.342759 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:12:53 crc kubenswrapper[4770]: I1209 12:12:53.340575 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:12:53 crc kubenswrapper[4770]: E1209 12:12:53.341483 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:13:07 crc kubenswrapper[4770]: I1209 12:13:07.340678 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:13:07 crc kubenswrapper[4770]: E1209 12:13:07.342498 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:13:20 crc kubenswrapper[4770]: I1209 12:13:20.340127 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:13:20 crc kubenswrapper[4770]: E1209 12:13:20.340840 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:13:35 crc kubenswrapper[4770]: I1209 12:13:35.382269 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:13:35 crc kubenswrapper[4770]: E1209 12:13:35.383141 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:13:49 crc kubenswrapper[4770]: I1209 12:13:49.343120 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:13:49 crc kubenswrapper[4770]: E1209 12:13:49.344180 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:00 crc kubenswrapper[4770]: I1209 12:14:00.340972 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:00 crc kubenswrapper[4770]: E1209 12:14:00.342796 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:13 crc kubenswrapper[4770]: I1209 12:14:13.340712 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:13 crc kubenswrapper[4770]: E1209 12:14:13.342149 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:24 crc kubenswrapper[4770]: I1209 12:14:24.341234 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:24 crc kubenswrapper[4770]: E1209 12:14:24.342713 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:35 crc kubenswrapper[4770]: I1209 12:14:35.344254 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:35 crc kubenswrapper[4770]: E1209 12:14:35.345044 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:47 crc kubenswrapper[4770]: I1209 12:14:47.340454 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:47 crc kubenswrapper[4770]: E1209 12:14:47.341229 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:14:59 crc kubenswrapper[4770]: I1209 12:14:59.341359 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:14:59 crc kubenswrapper[4770]: E1209 12:14:59.342563 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.155438 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf"] Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156120 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="extract-utilities" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156183 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="extract-utilities" Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156224 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="extract-content" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156236 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="extract-content" Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156257 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="extract-content" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156269 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="extract-content" Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156305 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="extract-utilities" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156327 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="extract-utilities" Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156355 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156365 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: E1209 12:15:00.156386 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156396 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156683 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cfe03e-73b5-4639-a892-4659f5212936" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.156706 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8d9807-4248-4897-a1e0-c62c6e27b38e" containerName="registry-server" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.160506 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.168848 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.173100 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.202408 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf"] Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.275831 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.275973 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjb4\" (UniqueName: \"kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.276021 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.377754 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.378662 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjb4\" (UniqueName: \"kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.378776 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.379042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.386367 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.400434 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjb4\" (UniqueName: \"kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4\") pod \"collect-profiles-29421375-xzvhf\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.504858 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:00 crc kubenswrapper[4770]: I1209 12:15:00.961356 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf"] Dec 09 12:15:01 crc kubenswrapper[4770]: I1209 12:15:01.950565 4770 generic.go:334] "Generic (PLEG): container finished" podID="160d1504-1e1a-4ec7-a52e-a253daa64800" containerID="b52d8c309e8d1fe31c53202ffad188bda04a5bec0f609f83ccdaa92450a477a4" exitCode=0 Dec 09 12:15:01 crc kubenswrapper[4770]: I1209 12:15:01.950667 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" event={"ID":"160d1504-1e1a-4ec7-a52e-a253daa64800","Type":"ContainerDied","Data":"b52d8c309e8d1fe31c53202ffad188bda04a5bec0f609f83ccdaa92450a477a4"} Dec 09 12:15:01 crc kubenswrapper[4770]: I1209 12:15:01.950891 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" event={"ID":"160d1504-1e1a-4ec7-a52e-a253daa64800","Type":"ContainerStarted","Data":"13dd83d19973b3b3acb7146b2f4c818924b8edec80fba73fa882b6ea30dc839d"} Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.233021 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.326151 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume\") pod \"160d1504-1e1a-4ec7-a52e-a253daa64800\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.326218 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnjb4\" (UniqueName: \"kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4\") pod \"160d1504-1e1a-4ec7-a52e-a253daa64800\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.326279 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume\") pod \"160d1504-1e1a-4ec7-a52e-a253daa64800\" (UID: \"160d1504-1e1a-4ec7-a52e-a253daa64800\") " Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.327051 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume" (OuterVolumeSpecName: "config-volume") pod "160d1504-1e1a-4ec7-a52e-a253daa64800" (UID: "160d1504-1e1a-4ec7-a52e-a253daa64800"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.333129 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4" (OuterVolumeSpecName: "kube-api-access-lnjb4") pod "160d1504-1e1a-4ec7-a52e-a253daa64800" (UID: "160d1504-1e1a-4ec7-a52e-a253daa64800"). InnerVolumeSpecName "kube-api-access-lnjb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.333397 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "160d1504-1e1a-4ec7-a52e-a253daa64800" (UID: "160d1504-1e1a-4ec7-a52e-a253daa64800"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.427848 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/160d1504-1e1a-4ec7-a52e-a253daa64800-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.427891 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnjb4\" (UniqueName: \"kubernetes.io/projected/160d1504-1e1a-4ec7-a52e-a253daa64800-kube-api-access-lnjb4\") on node \"crc\" DevicePath \"\"" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.427921 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/160d1504-1e1a-4ec7-a52e-a253daa64800-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.968308 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" event={"ID":"160d1504-1e1a-4ec7-a52e-a253daa64800","Type":"ContainerDied","Data":"13dd83d19973b3b3acb7146b2f4c818924b8edec80fba73fa882b6ea30dc839d"} Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.968376 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421375-xzvhf" Dec 09 12:15:03 crc kubenswrapper[4770]: I1209 12:15:03.968378 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13dd83d19973b3b3acb7146b2f4c818924b8edec80fba73fa882b6ea30dc839d" Dec 09 12:15:04 crc kubenswrapper[4770]: I1209 12:15:04.322786 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9"] Dec 09 12:15:04 crc kubenswrapper[4770]: I1209 12:15:04.327482 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421330-6vhh9"] Dec 09 12:15:05 crc kubenswrapper[4770]: I1209 12:15:05.348336 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f" path="/var/lib/kubelet/pods/0f2ca92c-3dd4-4ed7-9dad-bb80bc8b0b9f/volumes" Dec 09 12:15:14 crc kubenswrapper[4770]: I1209 12:15:14.340633 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:15:15 crc kubenswrapper[4770]: I1209 12:15:15.196142 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8"} Dec 09 12:15:30 crc kubenswrapper[4770]: I1209 12:15:30.911740 4770 scope.go:117] "RemoveContainer" containerID="0e3e6a8e0c829999802fedb626eb466ce74bf3d1effde6a72f83473af697eab3" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.061669 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tr6q"] Dec 09 12:16:15 crc kubenswrapper[4770]: E1209 12:16:15.062628 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160d1504-1e1a-4ec7-a52e-a253daa64800" containerName="collect-profiles" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.062642 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="160d1504-1e1a-4ec7-a52e-a253daa64800" containerName="collect-profiles" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.062824 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="160d1504-1e1a-4ec7-a52e-a253daa64800" containerName="collect-profiles" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.064024 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.069172 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tr6q"] Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.228308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcr2w\" (UniqueName: \"kubernetes.io/projected/0842ce6d-5d86-4008-ad0a-78674b554630-kube-api-access-wcr2w\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.228369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-utilities\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.228409 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-catalog-content\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.330810 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcr2w\" (UniqueName: \"kubernetes.io/projected/0842ce6d-5d86-4008-ad0a-78674b554630-kube-api-access-wcr2w\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.331121 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-utilities\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.331166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-catalog-content\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.331735 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-utilities\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.331759 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0842ce6d-5d86-4008-ad0a-78674b554630-catalog-content\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.358274 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcr2w\" (UniqueName: \"kubernetes.io/projected/0842ce6d-5d86-4008-ad0a-78674b554630-kube-api-access-wcr2w\") pod \"redhat-operators-7tr6q\" (UID: \"0842ce6d-5d86-4008-ad0a-78674b554630\") " pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.400932 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.892040 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tr6q"] Dec 09 12:16:15 crc kubenswrapper[4770]: I1209 12:16:15.935643 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tr6q" event={"ID":"0842ce6d-5d86-4008-ad0a-78674b554630","Type":"ContainerStarted","Data":"06f1dd934feb0c167ec70c5bfaafc5f40316f3405b223c26043d08b60f837e28"} Dec 09 12:16:16 crc kubenswrapper[4770]: I1209 12:16:16.947493 4770 generic.go:334] "Generic (PLEG): container finished" podID="0842ce6d-5d86-4008-ad0a-78674b554630" containerID="b69ffe622ef582fd51b69cb9a59a88112bd09f00f7732b5e799f7e0277f5c117" exitCode=0 Dec 09 12:16:16 crc kubenswrapper[4770]: I1209 12:16:16.947626 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tr6q" event={"ID":"0842ce6d-5d86-4008-ad0a-78674b554630","Type":"ContainerDied","Data":"b69ffe622ef582fd51b69cb9a59a88112bd09f00f7732b5e799f7e0277f5c117"} Dec 09 12:16:16 crc kubenswrapper[4770]: I1209 12:16:16.949605 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 12:16:25 crc kubenswrapper[4770]: I1209 12:16:25.131881 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tr6q" event={"ID":"0842ce6d-5d86-4008-ad0a-78674b554630","Type":"ContainerStarted","Data":"e74b1f8a3ea544763087d68f26cf795d70af5d8430f54f75f6ef18e35d2fbfaf"} Dec 09 12:16:26 crc kubenswrapper[4770]: I1209 12:16:26.141760 4770 generic.go:334] "Generic (PLEG): container finished" podID="0842ce6d-5d86-4008-ad0a-78674b554630" containerID="e74b1f8a3ea544763087d68f26cf795d70af5d8430f54f75f6ef18e35d2fbfaf" exitCode=0 Dec 09 12:16:26 crc kubenswrapper[4770]: I1209 12:16:26.141818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tr6q" event={"ID":"0842ce6d-5d86-4008-ad0a-78674b554630","Type":"ContainerDied","Data":"e74b1f8a3ea544763087d68f26cf795d70af5d8430f54f75f6ef18e35d2fbfaf"} Dec 09 12:16:27 crc kubenswrapper[4770]: I1209 12:16:27.232771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tr6q" event={"ID":"0842ce6d-5d86-4008-ad0a-78674b554630","Type":"ContainerStarted","Data":"ed431f8ce8b32c376b943eefb3d324c429630acadf863e848667440c193099f3"} Dec 09 12:16:27 crc kubenswrapper[4770]: I1209 12:16:27.260418 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tr6q" podStartSLOduration=2.657939399 podStartE2EDuration="12.260382588s" podCreationTimestamp="2025-12-09 12:16:15 +0000 UTC" firstStartedPulling="2025-12-09 12:16:16.949220107 +0000 UTC m=+2702.189978626" lastFinishedPulling="2025-12-09 12:16:26.551663296 +0000 UTC m=+2711.792421815" observedRunningTime="2025-12-09 12:16:27.248761925 +0000 UTC m=+2712.489520454" watchObservedRunningTime="2025-12-09 12:16:27.260382588 +0000 UTC m=+2712.501141107" Dec 09 12:16:35 crc kubenswrapper[4770]: I1209 12:16:35.401194 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:35 crc kubenswrapper[4770]: I1209 12:16:35.401747 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:35 crc kubenswrapper[4770]: I1209 12:16:35.448376 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:36 crc kubenswrapper[4770]: I1209 12:16:36.349483 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tr6q" Dec 09 12:16:36 crc kubenswrapper[4770]: I1209 12:16:36.434440 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tr6q"] Dec 09 12:16:36 crc kubenswrapper[4770]: I1209 12:16:36.469823 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 12:16:36 crc kubenswrapper[4770]: I1209 12:16:36.470155 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-twvqx" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="registry-server" containerID="cri-o://d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" gracePeriod=2 Dec 09 12:16:39 crc kubenswrapper[4770]: E1209 12:16:39.340772 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de is running failed: container process not found" containerID="d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 12:16:39 crc kubenswrapper[4770]: E1209 12:16:39.341741 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de is running failed: container process not found" containerID="d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 12:16:39 crc kubenswrapper[4770]: E1209 12:16:39.342149 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de is running failed: container process not found" containerID="d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 12:16:39 crc kubenswrapper[4770]: E1209 12:16:39.342202 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-twvqx" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="registry-server" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.162966 4770 generic.go:334] "Generic (PLEG): container finished" podID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerID="d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" exitCode=0 Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.163066 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerDied","Data":"d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de"} Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.447016 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.631433 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities\") pod \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.631798 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r585\" (UniqueName: \"kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585\") pod \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.631851 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content\") pod \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\" (UID: \"d7ea1b87-f78d-4999-a66a-3ca61f520d42\") " Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.632513 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities" (OuterVolumeSpecName: "utilities") pod "d7ea1b87-f78d-4999-a66a-3ca61f520d42" (UID: "d7ea1b87-f78d-4999-a66a-3ca61f520d42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.638432 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585" (OuterVolumeSpecName: "kube-api-access-2r585") pod "d7ea1b87-f78d-4999-a66a-3ca61f520d42" (UID: "d7ea1b87-f78d-4999-a66a-3ca61f520d42"). InnerVolumeSpecName "kube-api-access-2r585". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.643556 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.643603 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r585\" (UniqueName: \"kubernetes.io/projected/d7ea1b87-f78d-4999-a66a-3ca61f520d42-kube-api-access-2r585\") on node \"crc\" DevicePath \"\"" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.732361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7ea1b87-f78d-4999-a66a-3ca61f520d42" (UID: "d7ea1b87-f78d-4999-a66a-3ca61f520d42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:16:42 crc kubenswrapper[4770]: I1209 12:16:42.745318 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ea1b87-f78d-4999-a66a-3ca61f520d42-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.175430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twvqx" event={"ID":"d7ea1b87-f78d-4999-a66a-3ca61f520d42","Type":"ContainerDied","Data":"ad284524dbb643e747b7cbe6283a57f0fceb6b6a61f46c6111c3e5b5a18d5fc4"} Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.175514 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twvqx" Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.175543 4770 scope.go:117] "RemoveContainer" containerID="d07d808819648ab886484321d9c9e76902419e287d36b689d90c79391a2433de" Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.215565 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.217727 4770 scope.go:117] "RemoveContainer" containerID="7662d5b337f050813639e1ae72957acb9ad7eca1658c464f58447e4ec7d8a8dc" Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.224198 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-twvqx"] Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.241896 4770 scope.go:117] "RemoveContainer" containerID="c64d6bc91e0ec2cb3b9d127bd55fb64eab1fac535d0c0c43ce87c2958762b139" Dec 09 12:16:43 crc kubenswrapper[4770]: I1209 12:16:43.349608 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" path="/var/lib/kubelet/pods/d7ea1b87-f78d-4999-a66a-3ca61f520d42/volumes" Dec 09 12:17:32 crc kubenswrapper[4770]: I1209 12:17:32.474639 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:17:32 crc kubenswrapper[4770]: I1209 12:17:32.475573 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:18:02 crc kubenswrapper[4770]: I1209 12:18:02.474064 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:18:02 crc kubenswrapper[4770]: I1209 12:18:02.474890 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.669395 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:04 crc kubenswrapper[4770]: E1209 12:18:04.670451 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="extract-content" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.670485 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="extract-content" Dec 09 12:18:04 crc kubenswrapper[4770]: E1209 12:18:04.670507 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="extract-utilities" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.670538 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="extract-utilities" Dec 09 12:18:04 crc kubenswrapper[4770]: E1209 12:18:04.670553 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="registry-server" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.670560 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="registry-server" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.670957 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ea1b87-f78d-4999-a66a-3ca61f520d42" containerName="registry-server" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.673694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.685229 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.756475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vgfd\" (UniqueName: \"kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.756866 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.757072 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.858025 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vgfd\" (UniqueName: \"kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.858485 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.858688 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.859105 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.859164 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.879982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vgfd\" (UniqueName: \"kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd\") pod \"certified-operators-bkzmj\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:04 crc kubenswrapper[4770]: I1209 12:18:04.994314 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:05 crc kubenswrapper[4770]: I1209 12:18:05.762351 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:05 crc kubenswrapper[4770]: I1209 12:18:05.924378 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerStarted","Data":"8af86e93a8502dba28c685de9e58cca0340abacec4ab9a101832e284b84f15fe"} Dec 09 12:18:06 crc kubenswrapper[4770]: I1209 12:18:06.931810 4770 generic.go:334] "Generic (PLEG): container finished" podID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerID="7aa94c5f6e24dd75c32127f4c643abed21e6371969112669c44614bac06ec260" exitCode=0 Dec 09 12:18:06 crc kubenswrapper[4770]: I1209 12:18:06.931939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerDied","Data":"7aa94c5f6e24dd75c32127f4c643abed21e6371969112669c44614bac06ec260"} Dec 09 12:18:07 crc kubenswrapper[4770]: I1209 12:18:07.942246 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerStarted","Data":"b47984d0cc77bbdd614692c7c72f224c704f95253a5682b94b723870013f3c6c"} Dec 09 12:18:08 crc kubenswrapper[4770]: I1209 12:18:08.950816 4770 generic.go:334] "Generic (PLEG): container finished" podID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerID="b47984d0cc77bbdd614692c7c72f224c704f95253a5682b94b723870013f3c6c" exitCode=0 Dec 09 12:18:08 crc kubenswrapper[4770]: I1209 12:18:08.950917 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerDied","Data":"b47984d0cc77bbdd614692c7c72f224c704f95253a5682b94b723870013f3c6c"} Dec 09 12:18:10 crc kubenswrapper[4770]: I1209 12:18:10.006791 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerStarted","Data":"4a3a412f29f4000fe3a023d1f663eb93e8cfa23c8bb9507c92de96aca0f35abd"} Dec 09 12:18:10 crc kubenswrapper[4770]: I1209 12:18:10.028284 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bkzmj" podStartSLOduration=3.5574092999999998 podStartE2EDuration="6.02826576s" podCreationTimestamp="2025-12-09 12:18:04 +0000 UTC" firstStartedPulling="2025-12-09 12:18:06.934287261 +0000 UTC m=+2812.175045780" lastFinishedPulling="2025-12-09 12:18:09.405143721 +0000 UTC m=+2814.645902240" observedRunningTime="2025-12-09 12:18:10.023730445 +0000 UTC m=+2815.264488964" watchObservedRunningTime="2025-12-09 12:18:10.02826576 +0000 UTC m=+2815.269024279" Dec 09 12:18:14 crc kubenswrapper[4770]: I1209 12:18:14.994608 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:14 crc kubenswrapper[4770]: I1209 12:18:14.995060 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:15 crc kubenswrapper[4770]: I1209 12:18:15.069468 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:15 crc kubenswrapper[4770]: I1209 12:18:15.118741 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:15 crc kubenswrapper[4770]: I1209 12:18:15.330018 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:17 crc kubenswrapper[4770]: I1209 12:18:17.056518 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bkzmj" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="registry-server" containerID="cri-o://4a3a412f29f4000fe3a023d1f663eb93e8cfa23c8bb9507c92de96aca0f35abd" gracePeriod=2 Dec 09 12:18:18 crc kubenswrapper[4770]: I1209 12:18:18.069728 4770 generic.go:334] "Generic (PLEG): container finished" podID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerID="4a3a412f29f4000fe3a023d1f663eb93e8cfa23c8bb9507c92de96aca0f35abd" exitCode=0 Dec 09 12:18:18 crc kubenswrapper[4770]: I1209 12:18:18.069875 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerDied","Data":"4a3a412f29f4000fe3a023d1f663eb93e8cfa23c8bb9507c92de96aca0f35abd"} Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.847371 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.862732 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content\") pod \"07cc278a-95bb-488f-afce-9bfdd81246cd\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.862878 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities\") pod \"07cc278a-95bb-488f-afce-9bfdd81246cd\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.862927 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vgfd\" (UniqueName: \"kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd\") pod \"07cc278a-95bb-488f-afce-9bfdd81246cd\" (UID: \"07cc278a-95bb-488f-afce-9bfdd81246cd\") " Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.864346 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities" (OuterVolumeSpecName: "utilities") pod "07cc278a-95bb-488f-afce-9bfdd81246cd" (UID: "07cc278a-95bb-488f-afce-9bfdd81246cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.882156 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd" (OuterVolumeSpecName: "kube-api-access-8vgfd") pod "07cc278a-95bb-488f-afce-9bfdd81246cd" (UID: "07cc278a-95bb-488f-afce-9bfdd81246cd"). InnerVolumeSpecName "kube-api-access-8vgfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.931061 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07cc278a-95bb-488f-afce-9bfdd81246cd" (UID: "07cc278a-95bb-488f-afce-9bfdd81246cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.964240 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vgfd\" (UniqueName: \"kubernetes.io/projected/07cc278a-95bb-488f-afce-9bfdd81246cd-kube-api-access-8vgfd\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.964288 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:19 crc kubenswrapper[4770]: I1209 12:18:19.964298 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07cc278a-95bb-488f-afce-9bfdd81246cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.092966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkzmj" event={"ID":"07cc278a-95bb-488f-afce-9bfdd81246cd","Type":"ContainerDied","Data":"8af86e93a8502dba28c685de9e58cca0340abacec4ab9a101832e284b84f15fe"} Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.093059 4770 scope.go:117] "RemoveContainer" containerID="4a3a412f29f4000fe3a023d1f663eb93e8cfa23c8bb9507c92de96aca0f35abd" Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.093259 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkzmj" Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.132100 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.132321 4770 scope.go:117] "RemoveContainer" containerID="b47984d0cc77bbdd614692c7c72f224c704f95253a5682b94b723870013f3c6c" Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.140619 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bkzmj"] Dec 09 12:18:20 crc kubenswrapper[4770]: I1209 12:18:20.155120 4770 scope.go:117] "RemoveContainer" containerID="7aa94c5f6e24dd75c32127f4c643abed21e6371969112669c44614bac06ec260" Dec 09 12:18:21 crc kubenswrapper[4770]: I1209 12:18:21.351105 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" path="/var/lib/kubelet/pods/07cc278a-95bb-488f-afce-9bfdd81246cd/volumes" Dec 09 12:18:32 crc kubenswrapper[4770]: I1209 12:18:32.473817 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:18:32 crc kubenswrapper[4770]: I1209 12:18:32.474731 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:18:32 crc kubenswrapper[4770]: I1209 12:18:32.474801 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:18:32 crc kubenswrapper[4770]: I1209 12:18:32.475750 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:18:32 crc kubenswrapper[4770]: I1209 12:18:32.475820 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8" gracePeriod=600 Dec 09 12:18:33 crc kubenswrapper[4770]: I1209 12:18:33.199416 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8" exitCode=0 Dec 09 12:18:33 crc kubenswrapper[4770]: I1209 12:18:33.199491 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8"} Dec 09 12:18:33 crc kubenswrapper[4770]: I1209 12:18:33.199855 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111"} Dec 09 12:18:33 crc kubenswrapper[4770]: I1209 12:18:33.199880 4770 scope.go:117] "RemoveContainer" containerID="573d52c6f766a3f7eec8ad1da3e9680aaf6aaa63f59a9c137b1da4002c8f3a8a" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.692294 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:40 crc kubenswrapper[4770]: E1209 12:18:40.695752 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="extract-content" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.695776 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="extract-content" Dec 09 12:18:40 crc kubenswrapper[4770]: E1209 12:18:40.695793 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="extract-utilities" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.695799 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="extract-utilities" Dec 09 12:18:40 crc kubenswrapper[4770]: E1209 12:18:40.695814 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="registry-server" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.695820 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="registry-server" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.695981 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="07cc278a-95bb-488f-afce-9bfdd81246cd" containerName="registry-server" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.697191 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.706715 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.772308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.772382 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8rmv\" (UniqueName: \"kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.772406 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.873797 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.873962 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rmv\" (UniqueName: \"kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.873997 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.874497 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.874623 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:40 crc kubenswrapper[4770]: I1209 12:18:40.895057 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8rmv\" (UniqueName: \"kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv\") pod \"redhat-marketplace-mv96k\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:41 crc kubenswrapper[4770]: I1209 12:18:41.020351 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:41 crc kubenswrapper[4770]: I1209 12:18:41.601049 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:42 crc kubenswrapper[4770]: I1209 12:18:42.281659 4770 generic.go:334] "Generic (PLEG): container finished" podID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerID="8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689" exitCode=0 Dec 09 12:18:42 crc kubenswrapper[4770]: I1209 12:18:42.281924 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerDied","Data":"8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689"} Dec 09 12:18:42 crc kubenswrapper[4770]: I1209 12:18:42.281951 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerStarted","Data":"9b0ab8744206f35a390e90f9de986f6fdefc586a9c5d45708f5e6a85fc75650a"} Dec 09 12:18:43 crc kubenswrapper[4770]: I1209 12:18:43.290559 4770 generic.go:334] "Generic (PLEG): container finished" podID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerID="0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd" exitCode=0 Dec 09 12:18:43 crc kubenswrapper[4770]: I1209 12:18:43.290624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerDied","Data":"0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd"} Dec 09 12:18:44 crc kubenswrapper[4770]: I1209 12:18:44.305713 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerStarted","Data":"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124"} Dec 09 12:18:44 crc kubenswrapper[4770]: I1209 12:18:44.329647 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mv96k" podStartSLOduration=2.820176514 podStartE2EDuration="4.329627844s" podCreationTimestamp="2025-12-09 12:18:40 +0000 UTC" firstStartedPulling="2025-12-09 12:18:42.283221702 +0000 UTC m=+2847.523980221" lastFinishedPulling="2025-12-09 12:18:43.792673032 +0000 UTC m=+2849.033431551" observedRunningTime="2025-12-09 12:18:44.325823298 +0000 UTC m=+2849.566581817" watchObservedRunningTime="2025-12-09 12:18:44.329627844 +0000 UTC m=+2849.570386363" Dec 09 12:18:51 crc kubenswrapper[4770]: I1209 12:18:51.021530 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:51 crc kubenswrapper[4770]: I1209 12:18:51.022746 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:51 crc kubenswrapper[4770]: I1209 12:18:51.075043 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:51 crc kubenswrapper[4770]: I1209 12:18:51.463959 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:51 crc kubenswrapper[4770]: I1209 12:18:51.513887 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:53 crc kubenswrapper[4770]: I1209 12:18:53.422305 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mv96k" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="registry-server" containerID="cri-o://6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124" gracePeriod=2 Dec 09 12:18:53 crc kubenswrapper[4770]: I1209 12:18:53.897097 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.008023 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities\") pod \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.008148 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content\") pod \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.008289 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8rmv\" (UniqueName: \"kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv\") pod \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\" (UID: \"997c2bd4-08d9-40ea-bee1-a42c786a2b50\") " Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.008885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities" (OuterVolumeSpecName: "utilities") pod "997c2bd4-08d9-40ea-bee1-a42c786a2b50" (UID: "997c2bd4-08d9-40ea-bee1-a42c786a2b50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.014265 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv" (OuterVolumeSpecName: "kube-api-access-d8rmv") pod "997c2bd4-08d9-40ea-bee1-a42c786a2b50" (UID: "997c2bd4-08d9-40ea-bee1-a42c786a2b50"). InnerVolumeSpecName "kube-api-access-d8rmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.030232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "997c2bd4-08d9-40ea-bee1-a42c786a2b50" (UID: "997c2bd4-08d9-40ea-bee1-a42c786a2b50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.110185 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8rmv\" (UniqueName: \"kubernetes.io/projected/997c2bd4-08d9-40ea-bee1-a42c786a2b50-kube-api-access-d8rmv\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.110237 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.110252 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997c2bd4-08d9-40ea-bee1-a42c786a2b50-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.431609 4770 generic.go:334] "Generic (PLEG): container finished" podID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerID="6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124" exitCode=0 Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.431664 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerDied","Data":"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124"} Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.431677 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mv96k" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.431699 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mv96k" event={"ID":"997c2bd4-08d9-40ea-bee1-a42c786a2b50","Type":"ContainerDied","Data":"9b0ab8744206f35a390e90f9de986f6fdefc586a9c5d45708f5e6a85fc75650a"} Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.431719 4770 scope.go:117] "RemoveContainer" containerID="6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.454730 4770 scope.go:117] "RemoveContainer" containerID="0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.469038 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.472719 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mv96k"] Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.487749 4770 scope.go:117] "RemoveContainer" containerID="8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.506370 4770 scope.go:117] "RemoveContainer" containerID="6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124" Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.506889 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124\": container with ID starting with 6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124 not found: ID does not exist" containerID="6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.506949 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124"} err="failed to get container status \"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124\": rpc error: code = NotFound desc = could not find container \"6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124\": container with ID starting with 6a820af78abafeb3688a5549c9aeb4307858ebe20993c26c605421d3b7290124 not found: ID does not exist" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.506980 4770 scope.go:117] "RemoveContainer" containerID="0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd" Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.507269 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd\": container with ID starting with 0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd not found: ID does not exist" containerID="0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.507299 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd"} err="failed to get container status \"0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd\": rpc error: code = NotFound desc = could not find container \"0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd\": container with ID starting with 0c8ef27fb4d2c2ba6e5c8be77f72b40dd8a3b2af7c5ed017b8f2b5d23f6f22bd not found: ID does not exist" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.507317 4770 scope.go:117] "RemoveContainer" containerID="8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689" Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.507551 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689\": container with ID starting with 8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689 not found: ID does not exist" containerID="8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.507581 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689"} err="failed to get container status \"8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689\": rpc error: code = NotFound desc = could not find container \"8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689\": container with ID starting with 8c09a1a5cc727a993a984b8d936fedb584bed8eada984a2adfb4033f65d01689 not found: ID does not exist" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.730171 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.730601 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="extract-content" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.730623 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="extract-content" Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.730638 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="registry-server" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.730646 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="registry-server" Dec 09 12:18:54 crc kubenswrapper[4770]: E1209 12:18:54.730667 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="extract-utilities" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.730678 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="extract-utilities" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.730832 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" containerName="registry-server" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.732132 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.758177 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.821917 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgs99\" (UniqueName: \"kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.822308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.822481 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.923691 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgs99\" (UniqueName: \"kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.923765 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.923806 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.924528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.924960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:54 crc kubenswrapper[4770]: I1209 12:18:54.964912 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgs99\" (UniqueName: \"kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99\") pod \"community-operators-gg7f8\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:55 crc kubenswrapper[4770]: I1209 12:18:55.164037 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:18:55 crc kubenswrapper[4770]: I1209 12:18:55.364579 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="997c2bd4-08d9-40ea-bee1-a42c786a2b50" path="/var/lib/kubelet/pods/997c2bd4-08d9-40ea-bee1-a42c786a2b50/volumes" Dec 09 12:18:55 crc kubenswrapper[4770]: I1209 12:18:55.833786 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:18:55 crc kubenswrapper[4770]: W1209 12:18:55.844096 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0020246d_ea0b_4e9c_b441_95573bcaeb15.slice/crio-c9e47c2052b4b7e21acd8e2106f0d2756cf646aa7c9d2a60299dcc100f07bdff WatchSource:0}: Error finding container c9e47c2052b4b7e21acd8e2106f0d2756cf646aa7c9d2a60299dcc100f07bdff: Status 404 returned error can't find the container with id c9e47c2052b4b7e21acd8e2106f0d2756cf646aa7c9d2a60299dcc100f07bdff Dec 09 12:18:56 crc kubenswrapper[4770]: I1209 12:18:56.462414 4770 generic.go:334] "Generic (PLEG): container finished" podID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerID="46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2" exitCode=0 Dec 09 12:18:56 crc kubenswrapper[4770]: I1209 12:18:56.462461 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerDied","Data":"46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2"} Dec 09 12:18:56 crc kubenswrapper[4770]: I1209 12:18:56.462498 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerStarted","Data":"c9e47c2052b4b7e21acd8e2106f0d2756cf646aa7c9d2a60299dcc100f07bdff"} Dec 09 12:18:58 crc kubenswrapper[4770]: I1209 12:18:58.479390 4770 generic.go:334] "Generic (PLEG): container finished" podID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerID="d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705" exitCode=0 Dec 09 12:18:58 crc kubenswrapper[4770]: I1209 12:18:58.479494 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerDied","Data":"d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705"} Dec 09 12:18:59 crc kubenswrapper[4770]: I1209 12:18:59.490454 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerStarted","Data":"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55"} Dec 09 12:18:59 crc kubenswrapper[4770]: I1209 12:18:59.514712 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gg7f8" podStartSLOduration=3.026675048 podStartE2EDuration="5.514678011s" podCreationTimestamp="2025-12-09 12:18:54 +0000 UTC" firstStartedPulling="2025-12-09 12:18:56.464002616 +0000 UTC m=+2861.704761135" lastFinishedPulling="2025-12-09 12:18:58.952005579 +0000 UTC m=+2864.192764098" observedRunningTime="2025-12-09 12:18:59.510860414 +0000 UTC m=+2864.751618943" watchObservedRunningTime="2025-12-09 12:18:59.514678011 +0000 UTC m=+2864.755436530" Dec 09 12:19:05 crc kubenswrapper[4770]: I1209 12:19:05.271664 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:05 crc kubenswrapper[4770]: I1209 12:19:05.272332 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:05 crc kubenswrapper[4770]: I1209 12:19:05.327970 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:05 crc kubenswrapper[4770]: I1209 12:19:05.594241 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:05 crc kubenswrapper[4770]: I1209 12:19:05.653258 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:19:07 crc kubenswrapper[4770]: I1209 12:19:07.558791 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gg7f8" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="registry-server" containerID="cri-o://4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55" gracePeriod=2 Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.158709 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.315589 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities\") pod \"0020246d-ea0b-4e9c-b441-95573bcaeb15\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.315705 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content\") pod \"0020246d-ea0b-4e9c-b441-95573bcaeb15\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.315840 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgs99\" (UniqueName: \"kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99\") pod \"0020246d-ea0b-4e9c-b441-95573bcaeb15\" (UID: \"0020246d-ea0b-4e9c-b441-95573bcaeb15\") " Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.316707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities" (OuterVolumeSpecName: "utilities") pod "0020246d-ea0b-4e9c-b441-95573bcaeb15" (UID: "0020246d-ea0b-4e9c-b441-95573bcaeb15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.326256 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99" (OuterVolumeSpecName: "kube-api-access-qgs99") pod "0020246d-ea0b-4e9c-b441-95573bcaeb15" (UID: "0020246d-ea0b-4e9c-b441-95573bcaeb15"). InnerVolumeSpecName "kube-api-access-qgs99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.366380 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0020246d-ea0b-4e9c-b441-95573bcaeb15" (UID: "0020246d-ea0b-4e9c-b441-95573bcaeb15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.417731 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgs99\" (UniqueName: \"kubernetes.io/projected/0020246d-ea0b-4e9c-b441-95573bcaeb15-kube-api-access-qgs99\") on node \"crc\" DevicePath \"\"" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.417781 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.417792 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0020246d-ea0b-4e9c-b441-95573bcaeb15-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.575492 4770 generic.go:334] "Generic (PLEG): container finished" podID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerID="4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55" exitCode=0 Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.575541 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerDied","Data":"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55"} Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.575574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gg7f8" event={"ID":"0020246d-ea0b-4e9c-b441-95573bcaeb15","Type":"ContainerDied","Data":"c9e47c2052b4b7e21acd8e2106f0d2756cf646aa7c9d2a60299dcc100f07bdff"} Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.575576 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gg7f8" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.575595 4770 scope.go:117] "RemoveContainer" containerID="4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.596499 4770 scope.go:117] "RemoveContainer" containerID="d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.618716 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.624865 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gg7f8"] Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.634145 4770 scope.go:117] "RemoveContainer" containerID="46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.650817 4770 scope.go:117] "RemoveContainer" containerID="4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55" Dec 09 12:19:09 crc kubenswrapper[4770]: E1209 12:19:09.652030 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55\": container with ID starting with 4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55 not found: ID does not exist" containerID="4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.652074 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55"} err="failed to get container status \"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55\": rpc error: code = NotFound desc = could not find container \"4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55\": container with ID starting with 4ebe9da2d26dc57fa2be954f4304ea12fda5bfdb4b529b560e6e517dee9f2b55 not found: ID does not exist" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.652115 4770 scope.go:117] "RemoveContainer" containerID="d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705" Dec 09 12:19:09 crc kubenswrapper[4770]: E1209 12:19:09.652473 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705\": container with ID starting with d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705 not found: ID does not exist" containerID="d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.652516 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705"} err="failed to get container status \"d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705\": rpc error: code = NotFound desc = could not find container \"d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705\": container with ID starting with d04815f60586370a67888d78d2885b1c50b5520b5a10931a040b33123544f705 not found: ID does not exist" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.652533 4770 scope.go:117] "RemoveContainer" containerID="46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2" Dec 09 12:19:09 crc kubenswrapper[4770]: E1209 12:19:09.653069 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2\": container with ID starting with 46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2 not found: ID does not exist" containerID="46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2" Dec 09 12:19:09 crc kubenswrapper[4770]: I1209 12:19:09.653118 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2"} err="failed to get container status \"46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2\": rpc error: code = NotFound desc = could not find container \"46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2\": container with ID starting with 46d58154cc20da504af3f271c384ce6532cffbd206d4bad94b6a747c3f07cbc2 not found: ID does not exist" Dec 09 12:19:11 crc kubenswrapper[4770]: I1209 12:19:11.351053 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" path="/var/lib/kubelet/pods/0020246d-ea0b-4e9c-b441-95573bcaeb15/volumes" Dec 09 12:21:02 crc kubenswrapper[4770]: I1209 12:21:02.473971 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:21:02 crc kubenswrapper[4770]: I1209 12:21:02.474592 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:21:32 crc kubenswrapper[4770]: I1209 12:21:32.474319 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:21:32 crc kubenswrapper[4770]: I1209 12:21:32.474966 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:22:02 crc kubenswrapper[4770]: I1209 12:22:02.473713 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:22:02 crc kubenswrapper[4770]: I1209 12:22:02.474442 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:22:02 crc kubenswrapper[4770]: I1209 12:22:02.474539 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:22:02 crc kubenswrapper[4770]: I1209 12:22:02.475433 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:22:02 crc kubenswrapper[4770]: I1209 12:22:02.475524 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" gracePeriod=600 Dec 09 12:22:03 crc kubenswrapper[4770]: E1209 12:22:03.123838 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:22:03 crc kubenswrapper[4770]: I1209 12:22:03.481101 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" exitCode=0 Dec 09 12:22:03 crc kubenswrapper[4770]: I1209 12:22:03.481181 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111"} Dec 09 12:22:03 crc kubenswrapper[4770]: I1209 12:22:03.481282 4770 scope.go:117] "RemoveContainer" containerID="9ff64e7e6213faf41c2656210c56c2dd72e8c0d85ca7a003b502a8f102d89eb8" Dec 09 12:22:03 crc kubenswrapper[4770]: I1209 12:22:03.483835 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:22:03 crc kubenswrapper[4770]: E1209 12:22:03.484758 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:22:18 crc kubenswrapper[4770]: I1209 12:22:18.339790 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:22:18 crc kubenswrapper[4770]: E1209 12:22:18.340617 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:22:31 crc kubenswrapper[4770]: I1209 12:22:31.340474 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:22:31 crc kubenswrapper[4770]: E1209 12:22:31.341236 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:22:44 crc kubenswrapper[4770]: I1209 12:22:44.341084 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:22:44 crc kubenswrapper[4770]: E1209 12:22:44.341966 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:22:59 crc kubenswrapper[4770]: I1209 12:22:59.339716 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:22:59 crc kubenswrapper[4770]: E1209 12:22:59.340562 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:23:10 crc kubenswrapper[4770]: I1209 12:23:10.340457 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:23:10 crc kubenswrapper[4770]: E1209 12:23:10.341654 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:23:25 crc kubenswrapper[4770]: I1209 12:23:25.344744 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:23:25 crc kubenswrapper[4770]: E1209 12:23:25.345844 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:23:39 crc kubenswrapper[4770]: I1209 12:23:39.340009 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:23:39 crc kubenswrapper[4770]: E1209 12:23:39.341021 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:23:53 crc kubenswrapper[4770]: I1209 12:23:53.340765 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:23:53 crc kubenswrapper[4770]: E1209 12:23:53.341630 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:24:06 crc kubenswrapper[4770]: I1209 12:24:06.340613 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:24:06 crc kubenswrapper[4770]: E1209 12:24:06.341329 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:24:18 crc kubenswrapper[4770]: I1209 12:24:18.340356 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:24:18 crc kubenswrapper[4770]: E1209 12:24:18.341232 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:24:32 crc kubenswrapper[4770]: I1209 12:24:32.340788 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:24:32 crc kubenswrapper[4770]: E1209 12:24:32.341500 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:24:43 crc kubenswrapper[4770]: I1209 12:24:43.340023 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:24:43 crc kubenswrapper[4770]: E1209 12:24:43.340810 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:24:56 crc kubenswrapper[4770]: I1209 12:24:56.343067 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:24:56 crc kubenswrapper[4770]: E1209 12:24:56.344756 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:25:08 crc kubenswrapper[4770]: I1209 12:25:08.340610 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:25:08 crc kubenswrapper[4770]: E1209 12:25:08.341288 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:25:21 crc kubenswrapper[4770]: I1209 12:25:21.339989 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:25:21 crc kubenswrapper[4770]: E1209 12:25:21.340887 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:25:36 crc kubenswrapper[4770]: I1209 12:25:36.340316 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:25:36 crc kubenswrapper[4770]: E1209 12:25:36.341079 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:25:49 crc kubenswrapper[4770]: I1209 12:25:49.341770 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:25:49 crc kubenswrapper[4770]: E1209 12:25:49.342521 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:26:02 crc kubenswrapper[4770]: I1209 12:26:02.339967 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:26:02 crc kubenswrapper[4770]: E1209 12:26:02.340868 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:26:13 crc kubenswrapper[4770]: I1209 12:26:13.340141 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:26:13 crc kubenswrapper[4770]: E1209 12:26:13.340942 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:26:26 crc kubenswrapper[4770]: I1209 12:26:26.339770 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:26:26 crc kubenswrapper[4770]: E1209 12:26:26.340611 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:26:40 crc kubenswrapper[4770]: I1209 12:26:40.340811 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:26:40 crc kubenswrapper[4770]: E1209 12:26:40.341746 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:26:54 crc kubenswrapper[4770]: I1209 12:26:54.339880 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:26:54 crc kubenswrapper[4770]: E1209 12:26:54.340614 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:27:09 crc kubenswrapper[4770]: I1209 12:27:09.340669 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:27:11 crc kubenswrapper[4770]: I1209 12:27:11.450810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5"} Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.875733 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:28 crc kubenswrapper[4770]: E1209 12:28:28.876968 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="extract-content" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.876998 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="extract-content" Dec 09 12:28:28 crc kubenswrapper[4770]: E1209 12:28:28.877048 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="extract-utilities" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.877059 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="extract-utilities" Dec 09 12:28:28 crc kubenswrapper[4770]: E1209 12:28:28.877083 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="registry-server" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.877092 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="registry-server" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.877345 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0020246d-ea0b-4e9c-b441-95573bcaeb15" containerName="registry-server" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.880531 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.886457 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.986743 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.986828 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:28 crc kubenswrapper[4770]: I1209 12:28:28.987628 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmrsc\" (UniqueName: \"kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.089509 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.089598 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.089673 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmrsc\" (UniqueName: \"kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.090314 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.090517 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.112344 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmrsc\" (UniqueName: \"kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc\") pod \"certified-operators-dts2n\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.207563 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:29 crc kubenswrapper[4770]: I1209 12:28:29.549096 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:30 crc kubenswrapper[4770]: I1209 12:28:30.045088 4770 generic.go:334] "Generic (PLEG): container finished" podID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerID="d7cf3a1c5579dc8e3a0a870c525b090b2f288263673b72dd44ddd9d102c1628a" exitCode=0 Dec 09 12:28:30 crc kubenswrapper[4770]: I1209 12:28:30.045321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerDied","Data":"d7cf3a1c5579dc8e3a0a870c525b090b2f288263673b72dd44ddd9d102c1628a"} Dec 09 12:28:30 crc kubenswrapper[4770]: I1209 12:28:30.045446 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerStarted","Data":"4988cf14f86fd5c621e50abee5b9f28eba6d534434a81919cf1590de239b5ca8"} Dec 09 12:28:30 crc kubenswrapper[4770]: I1209 12:28:30.047740 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 12:28:31 crc kubenswrapper[4770]: I1209 12:28:31.060934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerStarted","Data":"d8d16f5083dc411fd1a5dca1b6c67531e0cb1506782305bee6a12febbed948de"} Dec 09 12:28:32 crc kubenswrapper[4770]: I1209 12:28:32.071828 4770 generic.go:334] "Generic (PLEG): container finished" podID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerID="d8d16f5083dc411fd1a5dca1b6c67531e0cb1506782305bee6a12febbed948de" exitCode=0 Dec 09 12:28:32 crc kubenswrapper[4770]: I1209 12:28:32.071935 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerDied","Data":"d8d16f5083dc411fd1a5dca1b6c67531e0cb1506782305bee6a12febbed948de"} Dec 09 12:28:34 crc kubenswrapper[4770]: I1209 12:28:34.093732 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerStarted","Data":"27da6f89a686cefa10cd82d4280c8ce7498546e70492cf397e2136fa92ee3bd1"} Dec 09 12:28:34 crc kubenswrapper[4770]: I1209 12:28:34.114457 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dts2n" podStartSLOduration=3.039014193 podStartE2EDuration="6.114425743s" podCreationTimestamp="2025-12-09 12:28:28 +0000 UTC" firstStartedPulling="2025-12-09 12:28:30.047328751 +0000 UTC m=+3435.288087270" lastFinishedPulling="2025-12-09 12:28:33.122740301 +0000 UTC m=+3438.363498820" observedRunningTime="2025-12-09 12:28:34.112761581 +0000 UTC m=+3439.353520110" watchObservedRunningTime="2025-12-09 12:28:34.114425743 +0000 UTC m=+3439.355184262" Dec 09 12:28:39 crc kubenswrapper[4770]: I1209 12:28:39.208721 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:39 crc kubenswrapper[4770]: I1209 12:28:39.210307 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:39 crc kubenswrapper[4770]: I1209 12:28:39.256405 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:40 crc kubenswrapper[4770]: I1209 12:28:40.194159 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:40 crc kubenswrapper[4770]: I1209 12:28:40.249792 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:42 crc kubenswrapper[4770]: I1209 12:28:42.160051 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dts2n" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="registry-server" containerID="cri-o://27da6f89a686cefa10cd82d4280c8ce7498546e70492cf397e2136fa92ee3bd1" gracePeriod=2 Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.172287 4770 generic.go:334] "Generic (PLEG): container finished" podID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerID="27da6f89a686cefa10cd82d4280c8ce7498546e70492cf397e2136fa92ee3bd1" exitCode=0 Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.172364 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerDied","Data":"27da6f89a686cefa10cd82d4280c8ce7498546e70492cf397e2136fa92ee3bd1"} Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.675381 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.848435 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities\") pod \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.848510 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmrsc\" (UniqueName: \"kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc\") pod \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.848675 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content\") pod \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\" (UID: \"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8\") " Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.849935 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities" (OuterVolumeSpecName: "utilities") pod "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" (UID: "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.855656 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc" (OuterVolumeSpecName: "kube-api-access-dmrsc") pod "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" (UID: "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8"). InnerVolumeSpecName "kube-api-access-dmrsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.901087 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" (UID: "ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.951078 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.951191 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:28:43 crc kubenswrapper[4770]: I1209 12:28:43.951208 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmrsc\" (UniqueName: \"kubernetes.io/projected/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8-kube-api-access-dmrsc\") on node \"crc\" DevicePath \"\"" Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.183393 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dts2n" event={"ID":"ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8","Type":"ContainerDied","Data":"4988cf14f86fd5c621e50abee5b9f28eba6d534434a81919cf1590de239b5ca8"} Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.183483 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dts2n" Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.184015 4770 scope.go:117] "RemoveContainer" containerID="27da6f89a686cefa10cd82d4280c8ce7498546e70492cf397e2136fa92ee3bd1" Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.217826 4770 scope.go:117] "RemoveContainer" containerID="d8d16f5083dc411fd1a5dca1b6c67531e0cb1506782305bee6a12febbed948de" Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.225358 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.232165 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dts2n"] Dec 09 12:28:44 crc kubenswrapper[4770]: I1209 12:28:44.248601 4770 scope.go:117] "RemoveContainer" containerID="d7cf3a1c5579dc8e3a0a870c525b090b2f288263673b72dd44ddd9d102c1628a" Dec 09 12:28:45 crc kubenswrapper[4770]: I1209 12:28:45.348737 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" path="/var/lib/kubelet/pods/ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8/volumes" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.841118 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:07 crc kubenswrapper[4770]: E1209 12:29:07.842924 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="registry-server" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.843024 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="registry-server" Dec 09 12:29:07 crc kubenswrapper[4770]: E1209 12:29:07.843094 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="extract-utilities" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.843157 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="extract-utilities" Dec 09 12:29:07 crc kubenswrapper[4770]: E1209 12:29:07.843235 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="extract-content" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.843292 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="extract-content" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.843574 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba95cc1c-3f83-4dcb-a7b4-22e03be1b2a8" containerName="registry-server" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.845814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:07 crc kubenswrapper[4770]: I1209 12:29:07.856796 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.038301 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.038859 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdcvn\" (UniqueName: \"kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.038947 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.139968 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.140035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdcvn\" (UniqueName: \"kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.140069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.140574 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.140842 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.181463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdcvn\" (UniqueName: \"kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn\") pod \"community-operators-l4z8b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.184292 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:08 crc kubenswrapper[4770]: I1209 12:29:08.783208 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:09 crc kubenswrapper[4770]: I1209 12:29:09.373251 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerID="b4c5649461abf8aa59c4dcfb3d5cedce94a23ec8ea3aa7d5049a84268622befa" exitCode=0 Dec 09 12:29:09 crc kubenswrapper[4770]: I1209 12:29:09.373312 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerDied","Data":"b4c5649461abf8aa59c4dcfb3d5cedce94a23ec8ea3aa7d5049a84268622befa"} Dec 09 12:29:09 crc kubenswrapper[4770]: I1209 12:29:09.373539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerStarted","Data":"ddcafb5c1089ea6c36c4af7fb351ef036e72998b009677d7a3c69b2a69ba9991"} Dec 09 12:29:10 crc kubenswrapper[4770]: I1209 12:29:10.383417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerStarted","Data":"b1a0d822d1fb915fb709b2c646464f31d5a7525a02a5d7722a9ec915e9aac9df"} Dec 09 12:29:11 crc kubenswrapper[4770]: I1209 12:29:11.393420 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerID="b1a0d822d1fb915fb709b2c646464f31d5a7525a02a5d7722a9ec915e9aac9df" exitCode=0 Dec 09 12:29:11 crc kubenswrapper[4770]: I1209 12:29:11.393659 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerDied","Data":"b1a0d822d1fb915fb709b2c646464f31d5a7525a02a5d7722a9ec915e9aac9df"} Dec 09 12:29:12 crc kubenswrapper[4770]: I1209 12:29:12.404051 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerStarted","Data":"64f69a2da662a58ec730dc2faf20744453a6296c0513f17edc9d80e157d1ba2a"} Dec 09 12:29:12 crc kubenswrapper[4770]: I1209 12:29:12.433023 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l4z8b" podStartSLOduration=2.870539645 podStartE2EDuration="5.432999866s" podCreationTimestamp="2025-12-09 12:29:07 +0000 UTC" firstStartedPulling="2025-12-09 12:29:09.37553425 +0000 UTC m=+3474.616292769" lastFinishedPulling="2025-12-09 12:29:11.937994471 +0000 UTC m=+3477.178752990" observedRunningTime="2025-12-09 12:29:12.425379633 +0000 UTC m=+3477.666138152" watchObservedRunningTime="2025-12-09 12:29:12.432999866 +0000 UTC m=+3477.673758385" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.239675 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.241813 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.254246 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.420221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dvzt\" (UniqueName: \"kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.420564 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.420833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.521870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.522380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dvzt\" (UniqueName: \"kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.522654 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.523210 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.523458 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.545106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dvzt\" (UniqueName: \"kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt\") pod \"redhat-marketplace-n9pzl\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:14 crc kubenswrapper[4770]: I1209 12:29:14.560379 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.318464 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.431889 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerStarted","Data":"1b3daf4c946843a03f312b8ea70193bad1b27bfe8718deaa30d96147500b59d5"} Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.637995 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.642751 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.649131 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.831750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkpsw\" (UniqueName: \"kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.832659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.832851 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.937471 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.937546 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.937690 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkpsw\" (UniqueName: \"kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.942380 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.942780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:15 crc kubenswrapper[4770]: I1209 12:29:15.963846 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkpsw\" (UniqueName: \"kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw\") pod \"redhat-operators-lgd9p\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:16 crc kubenswrapper[4770]: I1209 12:29:16.042107 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:16 crc kubenswrapper[4770]: I1209 12:29:16.370522 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:16 crc kubenswrapper[4770]: I1209 12:29:16.449415 4770 generic.go:334] "Generic (PLEG): container finished" podID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerID="c7041edde84836b80c14ce33c943aa54209b43cbfb17df816cceac45fa5f79de" exitCode=0 Dec 09 12:29:16 crc kubenswrapper[4770]: I1209 12:29:16.449637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerDied","Data":"c7041edde84836b80c14ce33c943aa54209b43cbfb17df816cceac45fa5f79de"} Dec 09 12:29:16 crc kubenswrapper[4770]: I1209 12:29:16.573191 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerStarted","Data":"84c653e5d3f108a19c1466e4653d636144f2d2f17dacba764564b9a687ffa830"} Dec 09 12:29:17 crc kubenswrapper[4770]: I1209 12:29:17.581560 4770 generic.go:334] "Generic (PLEG): container finished" podID="da6326de-8245-4b0e-b85a-9395e772e76c" containerID="40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33" exitCode=0 Dec 09 12:29:17 crc kubenswrapper[4770]: I1209 12:29:17.581621 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerDied","Data":"40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33"} Dec 09 12:29:18 crc kubenswrapper[4770]: I1209 12:29:18.185274 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:18 crc kubenswrapper[4770]: I1209 12:29:18.185602 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:18 crc kubenswrapper[4770]: I1209 12:29:18.233337 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:18 crc kubenswrapper[4770]: I1209 12:29:18.700735 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:19 crc kubenswrapper[4770]: I1209 12:29:19.664929 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerStarted","Data":"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3"} Dec 09 12:29:20 crc kubenswrapper[4770]: I1209 12:29:20.683428 4770 generic.go:334] "Generic (PLEG): container finished" podID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerID="efcab88bc12027e0158942e68156b3268665f3706fd470df489f35083e543f5e" exitCode=0 Dec 09 12:29:20 crc kubenswrapper[4770]: I1209 12:29:20.683528 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerDied","Data":"efcab88bc12027e0158942e68156b3268665f3706fd470df489f35083e543f5e"} Dec 09 12:29:21 crc kubenswrapper[4770]: I1209 12:29:21.828440 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:21 crc kubenswrapper[4770]: I1209 12:29:21.828666 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l4z8b" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="registry-server" containerID="cri-o://64f69a2da662a58ec730dc2faf20744453a6296c0513f17edc9d80e157d1ba2a" gracePeriod=2 Dec 09 12:29:22 crc kubenswrapper[4770]: I1209 12:29:22.733405 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerID="64f69a2da662a58ec730dc2faf20744453a6296c0513f17edc9d80e157d1ba2a" exitCode=0 Dec 09 12:29:22 crc kubenswrapper[4770]: I1209 12:29:22.734027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerDied","Data":"64f69a2da662a58ec730dc2faf20744453a6296c0513f17edc9d80e157d1ba2a"} Dec 09 12:29:22 crc kubenswrapper[4770]: I1209 12:29:22.740213 4770 generic.go:334] "Generic (PLEG): container finished" podID="da6326de-8245-4b0e-b85a-9395e772e76c" containerID="7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3" exitCode=0 Dec 09 12:29:22 crc kubenswrapper[4770]: I1209 12:29:22.740296 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerDied","Data":"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3"} Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.105481 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.122115 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities\") pod \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.122554 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdcvn\" (UniqueName: \"kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn\") pod \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.122734 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content\") pod \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\" (UID: \"ae9c582d-8bb5-4abe-844e-87cbff0de95b\") " Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.124135 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities" (OuterVolumeSpecName: "utilities") pod "ae9c582d-8bb5-4abe-844e-87cbff0de95b" (UID: "ae9c582d-8bb5-4abe-844e-87cbff0de95b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.130677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn" (OuterVolumeSpecName: "kube-api-access-jdcvn") pod "ae9c582d-8bb5-4abe-844e-87cbff0de95b" (UID: "ae9c582d-8bb5-4abe-844e-87cbff0de95b"). InnerVolumeSpecName "kube-api-access-jdcvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.203846 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae9c582d-8bb5-4abe-844e-87cbff0de95b" (UID: "ae9c582d-8bb5-4abe-844e-87cbff0de95b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.225757 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdcvn\" (UniqueName: \"kubernetes.io/projected/ae9c582d-8bb5-4abe-844e-87cbff0de95b-kube-api-access-jdcvn\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.225795 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.225805 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9c582d-8bb5-4abe-844e-87cbff0de95b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.752441 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerStarted","Data":"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f"} Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.755323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4z8b" event={"ID":"ae9c582d-8bb5-4abe-844e-87cbff0de95b","Type":"ContainerDied","Data":"ddcafb5c1089ea6c36c4af7fb351ef036e72998b009677d7a3c69b2a69ba9991"} Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.755402 4770 scope.go:117] "RemoveContainer" containerID="64f69a2da662a58ec730dc2faf20744453a6296c0513f17edc9d80e157d1ba2a" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.755407 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4z8b" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.759429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerStarted","Data":"d850ec2b735dd0d180948a872e5350f9ba9f7de90c12d36e7fb7b716092c4433"} Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.773680 4770 scope.go:117] "RemoveContainer" containerID="b1a0d822d1fb915fb709b2c646464f31d5a7525a02a5d7722a9ec915e9aac9df" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.782466 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgd9p" podStartSLOduration=3.190744926 podStartE2EDuration="8.782440888s" podCreationTimestamp="2025-12-09 12:29:15 +0000 UTC" firstStartedPulling="2025-12-09 12:29:17.58406196 +0000 UTC m=+3482.824820489" lastFinishedPulling="2025-12-09 12:29:23.175757932 +0000 UTC m=+3488.416516451" observedRunningTime="2025-12-09 12:29:23.778520868 +0000 UTC m=+3489.019279407" watchObservedRunningTime="2025-12-09 12:29:23.782440888 +0000 UTC m=+3489.023199397" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.803731 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.808221 4770 scope.go:117] "RemoveContainer" containerID="b4c5649461abf8aa59c4dcfb3d5cedce94a23ec8ea3aa7d5049a84268622befa" Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.948432 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l4z8b"] Dec 09 12:29:23 crc kubenswrapper[4770]: I1209 12:29:23.953357 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n9pzl" podStartSLOduration=3.951791221 podStartE2EDuration="9.953328308s" podCreationTimestamp="2025-12-09 12:29:14 +0000 UTC" firstStartedPulling="2025-12-09 12:29:16.57309313 +0000 UTC m=+3481.813851649" lastFinishedPulling="2025-12-09 12:29:22.574630217 +0000 UTC m=+3487.815388736" observedRunningTime="2025-12-09 12:29:23.944207836 +0000 UTC m=+3489.184966365" watchObservedRunningTime="2025-12-09 12:29:23.953328308 +0000 UTC m=+3489.194086837" Dec 09 12:29:24 crc kubenswrapper[4770]: I1209 12:29:24.560791 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:24 crc kubenswrapper[4770]: I1209 12:29:24.561870 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:25 crc kubenswrapper[4770]: I1209 12:29:25.358046 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" path="/var/lib/kubelet/pods/ae9c582d-8bb5-4abe-844e-87cbff0de95b/volumes" Dec 09 12:29:25 crc kubenswrapper[4770]: I1209 12:29:25.620399 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-n9pzl" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="registry-server" probeResult="failure" output=< Dec 09 12:29:25 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 12:29:25 crc kubenswrapper[4770]: > Dec 09 12:29:26 crc kubenswrapper[4770]: I1209 12:29:26.081457 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:26 crc kubenswrapper[4770]: I1209 12:29:26.081503 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:27 crc kubenswrapper[4770]: I1209 12:29:27.146147 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lgd9p" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="registry-server" probeResult="failure" output=< Dec 09 12:29:27 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 12:29:27 crc kubenswrapper[4770]: > Dec 09 12:29:32 crc kubenswrapper[4770]: I1209 12:29:32.474079 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:29:32 crc kubenswrapper[4770]: I1209 12:29:32.474428 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:29:34 crc kubenswrapper[4770]: I1209 12:29:34.609370 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:34 crc kubenswrapper[4770]: I1209 12:29:34.712487 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:34 crc kubenswrapper[4770]: I1209 12:29:34.845706 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:35 crc kubenswrapper[4770]: I1209 12:29:35.862335 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n9pzl" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="registry-server" containerID="cri-o://d850ec2b735dd0d180948a872e5350f9ba9f7de90c12d36e7fb7b716092c4433" gracePeriod=2 Dec 09 12:29:36 crc kubenswrapper[4770]: I1209 12:29:36.087224 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:36 crc kubenswrapper[4770]: I1209 12:29:36.135329 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:37 crc kubenswrapper[4770]: I1209 12:29:37.245663 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:37 crc kubenswrapper[4770]: I1209 12:29:37.889551 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgd9p" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="registry-server" containerID="cri-o://9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f" gracePeriod=2 Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.425014 4770 generic.go:334] "Generic (PLEG): container finished" podID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerID="d850ec2b735dd0d180948a872e5350f9ba9f7de90c12d36e7fb7b716092c4433" exitCode=0 Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.425122 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerDied","Data":"d850ec2b735dd0d180948a872e5350f9ba9f7de90c12d36e7fb7b716092c4433"} Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.522172 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.685033 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dvzt\" (UniqueName: \"kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt\") pod \"1c415f14-2db8-4eac-adf8-61b0beafb68f\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.685412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities\") pod \"1c415f14-2db8-4eac-adf8-61b0beafb68f\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.685449 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content\") pod \"1c415f14-2db8-4eac-adf8-61b0beafb68f\" (UID: \"1c415f14-2db8-4eac-adf8-61b0beafb68f\") " Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.686390 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities" (OuterVolumeSpecName: "utilities") pod "1c415f14-2db8-4eac-adf8-61b0beafb68f" (UID: "1c415f14-2db8-4eac-adf8-61b0beafb68f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.691163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt" (OuterVolumeSpecName: "kube-api-access-8dvzt") pod "1c415f14-2db8-4eac-adf8-61b0beafb68f" (UID: "1c415f14-2db8-4eac-adf8-61b0beafb68f"). InnerVolumeSpecName "kube-api-access-8dvzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.715320 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c415f14-2db8-4eac-adf8-61b0beafb68f" (UID: "1c415f14-2db8-4eac-adf8-61b0beafb68f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.787646 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.788015 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c415f14-2db8-4eac-adf8-61b0beafb68f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:40 crc kubenswrapper[4770]: I1209 12:29:40.788089 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dvzt\" (UniqueName: \"kubernetes.io/projected/1c415f14-2db8-4eac-adf8-61b0beafb68f-kube-api-access-8dvzt\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.239412 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.296777 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkpsw\" (UniqueName: \"kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw\") pod \"da6326de-8245-4b0e-b85a-9395e772e76c\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.296903 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content\") pod \"da6326de-8245-4b0e-b85a-9395e772e76c\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.296953 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities\") pod \"da6326de-8245-4b0e-b85a-9395e772e76c\" (UID: \"da6326de-8245-4b0e-b85a-9395e772e76c\") " Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.298221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities" (OuterVolumeSpecName: "utilities") pod "da6326de-8245-4b0e-b85a-9395e772e76c" (UID: "da6326de-8245-4b0e-b85a-9395e772e76c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.302054 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw" (OuterVolumeSpecName: "kube-api-access-hkpsw") pod "da6326de-8245-4b0e-b85a-9395e772e76c" (UID: "da6326de-8245-4b0e-b85a-9395e772e76c"). InnerVolumeSpecName "kube-api-access-hkpsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.399225 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.399267 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkpsw\" (UniqueName: \"kubernetes.io/projected/da6326de-8245-4b0e-b85a-9395e772e76c-kube-api-access-hkpsw\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.429461 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da6326de-8245-4b0e-b85a-9395e772e76c" (UID: "da6326de-8245-4b0e-b85a-9395e772e76c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.466312 4770 generic.go:334] "Generic (PLEG): container finished" podID="da6326de-8245-4b0e-b85a-9395e772e76c" containerID="9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f" exitCode=0 Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.466394 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgd9p" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.466427 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerDied","Data":"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f"} Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.466462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgd9p" event={"ID":"da6326de-8245-4b0e-b85a-9395e772e76c","Type":"ContainerDied","Data":"84c653e5d3f108a19c1466e4653d636144f2d2f17dacba764564b9a687ffa830"} Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.466482 4770 scope.go:117] "RemoveContainer" containerID="9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.471237 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9pzl" event={"ID":"1c415f14-2db8-4eac-adf8-61b0beafb68f","Type":"ContainerDied","Data":"1b3daf4c946843a03f312b8ea70193bad1b27bfe8718deaa30d96147500b59d5"} Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.471346 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9pzl" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.500809 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6326de-8245-4b0e-b85a-9395e772e76c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.501224 4770 scope.go:117] "RemoveContainer" containerID="7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.502133 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.511407 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9pzl"] Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.520117 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.524469 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgd9p"] Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.525511 4770 scope.go:117] "RemoveContainer" containerID="40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.548200 4770 scope.go:117] "RemoveContainer" containerID="9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f" Dec 09 12:29:41 crc kubenswrapper[4770]: E1209 12:29:41.548765 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f\": container with ID starting with 9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f not found: ID does not exist" containerID="9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.548818 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f"} err="failed to get container status \"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f\": rpc error: code = NotFound desc = could not find container \"9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f\": container with ID starting with 9b20397743aced20b300c0e4a2eb23f2a14b8b42367836d697b0184015ce7e3f not found: ID does not exist" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.548864 4770 scope.go:117] "RemoveContainer" containerID="7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3" Dec 09 12:29:41 crc kubenswrapper[4770]: E1209 12:29:41.549308 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3\": container with ID starting with 7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3 not found: ID does not exist" containerID="7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.549349 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3"} err="failed to get container status \"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3\": rpc error: code = NotFound desc = could not find container \"7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3\": container with ID starting with 7fd00ccc3385a3ed6c409ca06d961d5847dab6db5525a2924365f3d7df3e05e3 not found: ID does not exist" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.549365 4770 scope.go:117] "RemoveContainer" containerID="40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33" Dec 09 12:29:41 crc kubenswrapper[4770]: E1209 12:29:41.549618 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33\": container with ID starting with 40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33 not found: ID does not exist" containerID="40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.549637 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33"} err="failed to get container status \"40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33\": rpc error: code = NotFound desc = could not find container \"40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33\": container with ID starting with 40be50a5b93a54b1dc9ebcdad1179ef261bed094bbe8dc247b1b1085f5092f33 not found: ID does not exist" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.549665 4770 scope.go:117] "RemoveContainer" containerID="d850ec2b735dd0d180948a872e5350f9ba9f7de90c12d36e7fb7b716092c4433" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.574809 4770 scope.go:117] "RemoveContainer" containerID="efcab88bc12027e0158942e68156b3268665f3706fd470df489f35083e543f5e" Dec 09 12:29:41 crc kubenswrapper[4770]: I1209 12:29:41.594014 4770 scope.go:117] "RemoveContainer" containerID="c7041edde84836b80c14ce33c943aa54209b43cbfb17df816cceac45fa5f79de" Dec 09 12:29:43 crc kubenswrapper[4770]: I1209 12:29:43.351063 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" path="/var/lib/kubelet/pods/1c415f14-2db8-4eac-adf8-61b0beafb68f/volumes" Dec 09 12:29:43 crc kubenswrapper[4770]: I1209 12:29:43.352062 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" path="/var/lib/kubelet/pods/da6326de-8245-4b0e-b85a-9395e772e76c/volumes" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.161592 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4"] Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163738 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163763 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163790 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163799 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163818 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163830 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163845 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163858 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163874 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163882 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163916 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.163931 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.163948 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164022 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.164047 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164055 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="extract-content" Dec 09 12:30:00 crc kubenswrapper[4770]: E1209 12:30:00.164068 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164075 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="extract-utilities" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164281 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c415f14-2db8-4eac-adf8-61b0beafb68f" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164308 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6326de-8245-4b0e-b85a-9395e772e76c" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.164328 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9c582d-8bb5-4abe-844e-87cbff0de95b" containerName="registry-server" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.165478 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.168242 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.168252 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.185183 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4"] Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.358944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h564m\" (UniqueName: \"kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.359408 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.359474 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.460488 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h564m\" (UniqueName: \"kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.460576 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.460607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.462392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.470986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.480134 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h564m\" (UniqueName: \"kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m\") pod \"collect-profiles-29421390-2zxs4\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.500007 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:00 crc kubenswrapper[4770]: I1209 12:30:00.945659 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4"] Dec 09 12:30:01 crc kubenswrapper[4770]: I1209 12:30:01.632161 4770 generic.go:334] "Generic (PLEG): container finished" podID="3ae7c8e5-6b1e-476b-9912-73566382be47" containerID="29b8aac28a35fda6ea1e1eed7b66450ebac2247b7c7e9bf9b3667e075f5c1e70" exitCode=0 Dec 09 12:30:01 crc kubenswrapper[4770]: I1209 12:30:01.632252 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" event={"ID":"3ae7c8e5-6b1e-476b-9912-73566382be47","Type":"ContainerDied","Data":"29b8aac28a35fda6ea1e1eed7b66450ebac2247b7c7e9bf9b3667e075f5c1e70"} Dec 09 12:30:01 crc kubenswrapper[4770]: I1209 12:30:01.632282 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" event={"ID":"3ae7c8e5-6b1e-476b-9912-73566382be47","Type":"ContainerStarted","Data":"e2c1baea0345da0d9972dbbb183f1ac33e4ef187a8c3881e974502568da3e6fa"} Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.473992 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.474341 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.892139 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.997756 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume\") pod \"3ae7c8e5-6b1e-476b-9912-73566382be47\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.997868 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume\") pod \"3ae7c8e5-6b1e-476b-9912-73566382be47\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.997947 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h564m\" (UniqueName: \"kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m\") pod \"3ae7c8e5-6b1e-476b-9912-73566382be47\" (UID: \"3ae7c8e5-6b1e-476b-9912-73566382be47\") " Dec 09 12:30:02 crc kubenswrapper[4770]: I1209 12:30:02.998973 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume" (OuterVolumeSpecName: "config-volume") pod "3ae7c8e5-6b1e-476b-9912-73566382be47" (UID: "3ae7c8e5-6b1e-476b-9912-73566382be47"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.005245 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3ae7c8e5-6b1e-476b-9912-73566382be47" (UID: "3ae7c8e5-6b1e-476b-9912-73566382be47"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.012105 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m" (OuterVolumeSpecName: "kube-api-access-h564m") pod "3ae7c8e5-6b1e-476b-9912-73566382be47" (UID: "3ae7c8e5-6b1e-476b-9912-73566382be47"). InnerVolumeSpecName "kube-api-access-h564m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.099346 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ae7c8e5-6b1e-476b-9912-73566382be47-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.099389 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ae7c8e5-6b1e-476b-9912-73566382be47-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.099401 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h564m\" (UniqueName: \"kubernetes.io/projected/3ae7c8e5-6b1e-476b-9912-73566382be47-kube-api-access-h564m\") on node \"crc\" DevicePath \"\"" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.657942 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" event={"ID":"3ae7c8e5-6b1e-476b-9912-73566382be47","Type":"ContainerDied","Data":"e2c1baea0345da0d9972dbbb183f1ac33e4ef187a8c3881e974502568da3e6fa"} Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.658306 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2c1baea0345da0d9972dbbb183f1ac33e4ef187a8c3881e974502568da3e6fa" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.658033 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421390-2zxs4" Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.977242 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk"] Dec 09 12:30:03 crc kubenswrapper[4770]: I1209 12:30:03.984424 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421345-7stsk"] Dec 09 12:30:05 crc kubenswrapper[4770]: I1209 12:30:05.351029 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b81a4879-e6d0-4eb8-b823-93b45c1fb249" path="/var/lib/kubelet/pods/b81a4879-e6d0-4eb8-b823-93b45c1fb249/volumes" Dec 09 12:30:31 crc kubenswrapper[4770]: I1209 12:30:31.271593 4770 scope.go:117] "RemoveContainer" containerID="53d9de51c2168967fc90a6f74c3d89222c199ce1f88cd29a368f94acdaf27178" Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.474323 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.474652 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.474707 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.475442 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.475499 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5" gracePeriod=600 Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.880604 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5" exitCode=0 Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.880695 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5"} Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.881344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd"} Dec 09 12:30:32 crc kubenswrapper[4770]: I1209 12:30:32.881407 4770 scope.go:117] "RemoveContainer" containerID="98a6b26ab360c819b8c4b6f76f168d34c498d8d45e7eaddade9f87166aa12111" Dec 09 12:32:32 crc kubenswrapper[4770]: I1209 12:32:32.473822 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:32:32 crc kubenswrapper[4770]: I1209 12:32:32.474452 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:33:02 crc kubenswrapper[4770]: I1209 12:33:02.473163 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:33:02 crc kubenswrapper[4770]: I1209 12:33:02.473673 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:33:32 crc kubenswrapper[4770]: I1209 12:33:32.473144 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:33:32 crc kubenswrapper[4770]: I1209 12:33:32.473643 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:33:32 crc kubenswrapper[4770]: I1209 12:33:32.473695 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:33:32 crc kubenswrapper[4770]: I1209 12:33:32.474357 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:33:32 crc kubenswrapper[4770]: I1209 12:33:32.474423 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" gracePeriod=600 Dec 09 12:33:32 crc kubenswrapper[4770]: E1209 12:33:32.595140 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:33:33 crc kubenswrapper[4770]: I1209 12:33:33.289743 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" exitCode=0 Dec 09 12:33:33 crc kubenswrapper[4770]: I1209 12:33:33.289812 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd"} Dec 09 12:33:33 crc kubenswrapper[4770]: I1209 12:33:33.290145 4770 scope.go:117] "RemoveContainer" containerID="6dbeaf8ae578782d83d28eafa54c8b00207b57af3ac7c5559a2c8baadd6ab7f5" Dec 09 12:33:33 crc kubenswrapper[4770]: I1209 12:33:33.290634 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:33:33 crc kubenswrapper[4770]: E1209 12:33:33.290868 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:33:47 crc kubenswrapper[4770]: I1209 12:33:47.340714 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:33:47 crc kubenswrapper[4770]: E1209 12:33:47.341494 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:33:59 crc kubenswrapper[4770]: I1209 12:33:59.339850 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:33:59 crc kubenswrapper[4770]: E1209 12:33:59.340691 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:34:11 crc kubenswrapper[4770]: I1209 12:34:11.340357 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:34:11 crc kubenswrapper[4770]: E1209 12:34:11.342228 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:34:22 crc kubenswrapper[4770]: I1209 12:34:22.341335 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:34:22 crc kubenswrapper[4770]: E1209 12:34:22.342012 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:34:35 crc kubenswrapper[4770]: I1209 12:34:35.344333 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:34:35 crc kubenswrapper[4770]: E1209 12:34:35.345019 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:34:47 crc kubenswrapper[4770]: I1209 12:34:47.340463 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:34:47 crc kubenswrapper[4770]: E1209 12:34:47.341420 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:01 crc kubenswrapper[4770]: I1209 12:35:01.340502 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:01 crc kubenswrapper[4770]: E1209 12:35:01.341265 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:14 crc kubenswrapper[4770]: I1209 12:35:14.340676 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:14 crc kubenswrapper[4770]: E1209 12:35:14.341482 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:27 crc kubenswrapper[4770]: I1209 12:35:27.340587 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:27 crc kubenswrapper[4770]: E1209 12:35:27.341312 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:38 crc kubenswrapper[4770]: I1209 12:35:38.339981 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:38 crc kubenswrapper[4770]: E1209 12:35:38.340723 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:48 crc kubenswrapper[4770]: I1209 12:35:48.473418 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:48 crc kubenswrapper[4770]: E1209 12:35:48.474674 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:35:59 crc kubenswrapper[4770]: I1209 12:35:59.340422 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:35:59 crc kubenswrapper[4770]: E1209 12:35:59.341618 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:36:11 crc kubenswrapper[4770]: I1209 12:36:11.340659 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:36:11 crc kubenswrapper[4770]: E1209 12:36:11.341457 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:36:25 crc kubenswrapper[4770]: I1209 12:36:25.344544 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:36:25 crc kubenswrapper[4770]: E1209 12:36:25.346306 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:36:37 crc kubenswrapper[4770]: I1209 12:36:37.340860 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:36:37 crc kubenswrapper[4770]: E1209 12:36:37.341779 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:36:48 crc kubenswrapper[4770]: I1209 12:36:48.339997 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:36:48 crc kubenswrapper[4770]: E1209 12:36:48.340670 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:37:03 crc kubenswrapper[4770]: I1209 12:37:03.340499 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:37:03 crc kubenswrapper[4770]: E1209 12:37:03.341417 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:37:17 crc kubenswrapper[4770]: I1209 12:37:17.344650 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:37:17 crc kubenswrapper[4770]: E1209 12:37:17.345609 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:37:31 crc kubenswrapper[4770]: I1209 12:37:31.364309 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:37:31 crc kubenswrapper[4770]: E1209 12:37:31.367264 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:37:42 crc kubenswrapper[4770]: I1209 12:37:42.340719 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:37:42 crc kubenswrapper[4770]: E1209 12:37:42.341334 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:37:57 crc kubenswrapper[4770]: I1209 12:37:57.342676 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:37:57 crc kubenswrapper[4770]: E1209 12:37:57.343645 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:38:11 crc kubenswrapper[4770]: I1209 12:38:11.339875 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:38:11 crc kubenswrapper[4770]: E1209 12:38:11.340975 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:38:26 crc kubenswrapper[4770]: I1209 12:38:26.340381 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:38:26 crc kubenswrapper[4770]: E1209 12:38:26.341102 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:38:39 crc kubenswrapper[4770]: I1209 12:38:39.340273 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:38:40 crc kubenswrapper[4770]: I1209 12:38:40.219308 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d"} Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.327122 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:38:46 crc kubenswrapper[4770]: E1209 12:38:46.328210 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae7c8e5-6b1e-476b-9912-73566382be47" containerName="collect-profiles" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.328236 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae7c8e5-6b1e-476b-9912-73566382be47" containerName="collect-profiles" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.329192 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae7c8e5-6b1e-476b-9912-73566382be47" containerName="collect-profiles" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.331707 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.337553 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.342685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.342736 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.342913 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfhm\" (UniqueName: \"kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.444191 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.444260 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.444332 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfhm\" (UniqueName: \"kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.445074 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.445164 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.469193 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfhm\" (UniqueName: \"kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm\") pod \"certified-operators-qhcll\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:46 crc kubenswrapper[4770]: I1209 12:38:46.653011 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:47 crc kubenswrapper[4770]: I1209 12:38:47.156135 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:38:47 crc kubenswrapper[4770]: I1209 12:38:47.267286 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerStarted","Data":"cbfdcf1737930ef48351b25de2aa5db96cca55092fc604376d65f5f76357b750"} Dec 09 12:38:48 crc kubenswrapper[4770]: I1209 12:38:48.276538 4770 generic.go:334] "Generic (PLEG): container finished" podID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerID="73049d7d0a5df746c20c0150a49c70008c24ad00bdaef46d37de080a7e6bc503" exitCode=0 Dec 09 12:38:48 crc kubenswrapper[4770]: I1209 12:38:48.276598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerDied","Data":"73049d7d0a5df746c20c0150a49c70008c24ad00bdaef46d37de080a7e6bc503"} Dec 09 12:38:48 crc kubenswrapper[4770]: I1209 12:38:48.279231 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 12:38:50 crc kubenswrapper[4770]: I1209 12:38:50.294160 4770 generic.go:334] "Generic (PLEG): container finished" podID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerID="1791cf2d873e4c87c24969ac8c31312a0ec14dead476d6e8edba6ae8995f0b88" exitCode=0 Dec 09 12:38:50 crc kubenswrapper[4770]: I1209 12:38:50.294272 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerDied","Data":"1791cf2d873e4c87c24969ac8c31312a0ec14dead476d6e8edba6ae8995f0b88"} Dec 09 12:38:51 crc kubenswrapper[4770]: I1209 12:38:51.305674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerStarted","Data":"fef0e6a4f484094d2f269882d6ecd31be454742544a8fde5909294b9fa8883ff"} Dec 09 12:38:51 crc kubenswrapper[4770]: I1209 12:38:51.363935 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qhcll" podStartSLOduration=2.931326817 podStartE2EDuration="5.363894241s" podCreationTimestamp="2025-12-09 12:38:46 +0000 UTC" firstStartedPulling="2025-12-09 12:38:48.278855024 +0000 UTC m=+4053.519613543" lastFinishedPulling="2025-12-09 12:38:50.711422448 +0000 UTC m=+4055.952180967" observedRunningTime="2025-12-09 12:38:51.359789136 +0000 UTC m=+4056.600547655" watchObservedRunningTime="2025-12-09 12:38:51.363894241 +0000 UTC m=+4056.604652760" Dec 09 12:38:56 crc kubenswrapper[4770]: I1209 12:38:56.653743 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:56 crc kubenswrapper[4770]: I1209 12:38:56.654761 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:56 crc kubenswrapper[4770]: I1209 12:38:56.705565 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:57 crc kubenswrapper[4770]: I1209 12:38:57.392407 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:38:57 crc kubenswrapper[4770]: I1209 12:38:57.448573 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:39:00 crc kubenswrapper[4770]: I1209 12:38:59.364657 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qhcll" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="registry-server" containerID="cri-o://fef0e6a4f484094d2f269882d6ecd31be454742544a8fde5909294b9fa8883ff" gracePeriod=2 Dec 09 12:39:00 crc kubenswrapper[4770]: I1209 12:39:00.373326 4770 generic.go:334] "Generic (PLEG): container finished" podID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerID="fef0e6a4f484094d2f269882d6ecd31be454742544a8fde5909294b9fa8883ff" exitCode=0 Dec 09 12:39:00 crc kubenswrapper[4770]: I1209 12:39:00.373659 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerDied","Data":"fef0e6a4f484094d2f269882d6ecd31be454742544a8fde5909294b9fa8883ff"} Dec 09 12:39:00 crc kubenswrapper[4770]: I1209 12:39:00.930486 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.065105 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdfhm\" (UniqueName: \"kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm\") pod \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.065486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content\") pod \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.065737 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities\") pod \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\" (UID: \"b867d1e1-e6fa-446a-86ae-c5590e13bb7b\") " Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.066726 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities" (OuterVolumeSpecName: "utilities") pod "b867d1e1-e6fa-446a-86ae-c5590e13bb7b" (UID: "b867d1e1-e6fa-446a-86ae-c5590e13bb7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.072799 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm" (OuterVolumeSpecName: "kube-api-access-qdfhm") pod "b867d1e1-e6fa-446a-86ae-c5590e13bb7b" (UID: "b867d1e1-e6fa-446a-86ae-c5590e13bb7b"). InnerVolumeSpecName "kube-api-access-qdfhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.124500 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b867d1e1-e6fa-446a-86ae-c5590e13bb7b" (UID: "b867d1e1-e6fa-446a-86ae-c5590e13bb7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.167116 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdfhm\" (UniqueName: \"kubernetes.io/projected/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-kube-api-access-qdfhm\") on node \"crc\" DevicePath \"\"" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.167151 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.167161 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b867d1e1-e6fa-446a-86ae-c5590e13bb7b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.385862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhcll" event={"ID":"b867d1e1-e6fa-446a-86ae-c5590e13bb7b","Type":"ContainerDied","Data":"cbfdcf1737930ef48351b25de2aa5db96cca55092fc604376d65f5f76357b750"} Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.385938 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhcll" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.385949 4770 scope.go:117] "RemoveContainer" containerID="fef0e6a4f484094d2f269882d6ecd31be454742544a8fde5909294b9fa8883ff" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.413157 4770 scope.go:117] "RemoveContainer" containerID="1791cf2d873e4c87c24969ac8c31312a0ec14dead476d6e8edba6ae8995f0b88" Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.419311 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.427252 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qhcll"] Dec 09 12:39:01 crc kubenswrapper[4770]: I1209 12:39:01.441587 4770 scope.go:117] "RemoveContainer" containerID="73049d7d0a5df746c20c0150a49c70008c24ad00bdaef46d37de080a7e6bc503" Dec 09 12:39:03 crc kubenswrapper[4770]: I1209 12:39:03.350221 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" path="/var/lib/kubelet/pods/b867d1e1-e6fa-446a-86ae-c5590e13bb7b/volumes" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.008766 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:39:56 crc kubenswrapper[4770]: E1209 12:39:56.009777 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="extract-utilities" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.009797 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="extract-utilities" Dec 09 12:39:56 crc kubenswrapper[4770]: E1209 12:39:56.009840 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="registry-server" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.009849 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="registry-server" Dec 09 12:39:56 crc kubenswrapper[4770]: E1209 12:39:56.009864 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="extract-content" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.010299 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="extract-content" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.010517 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b867d1e1-e6fa-446a-86ae-c5590e13bb7b" containerName="registry-server" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.011924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.021609 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.115245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.115300 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.115405 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntknq\" (UniqueName: \"kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.216423 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntknq\" (UniqueName: \"kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.216857 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.217042 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.217730 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.217857 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.245872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntknq\" (UniqueName: \"kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq\") pod \"redhat-operators-57mzf\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.369619 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:39:56 crc kubenswrapper[4770]: I1209 12:39:56.718925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:39:57 crc kubenswrapper[4770]: I1209 12:39:57.120950 4770 generic.go:334] "Generic (PLEG): container finished" podID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerID="30a2fedba9430ce527d49f9bc9cc9679f048495e1c649f68e3113b451961e28b" exitCode=0 Dec 09 12:39:57 crc kubenswrapper[4770]: I1209 12:39:57.121005 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerDied","Data":"30a2fedba9430ce527d49f9bc9cc9679f048495e1c649f68e3113b451961e28b"} Dec 09 12:39:57 crc kubenswrapper[4770]: I1209 12:39:57.121275 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerStarted","Data":"4ab4aac0fdad7a15e99f495d5cf2cec15ea9f4a835502e63603356bd0b951aa8"} Dec 09 12:39:58 crc kubenswrapper[4770]: I1209 12:39:58.129522 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerStarted","Data":"d0628aee28c653c06a67e4d0a81a53c216405d3cf0036ea3002e56b56226a561"} Dec 09 12:40:00 crc kubenswrapper[4770]: I1209 12:40:00.146713 4770 generic.go:334] "Generic (PLEG): container finished" podID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerID="d0628aee28c653c06a67e4d0a81a53c216405d3cf0036ea3002e56b56226a561" exitCode=0 Dec 09 12:40:00 crc kubenswrapper[4770]: I1209 12:40:00.146809 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerDied","Data":"d0628aee28c653c06a67e4d0a81a53c216405d3cf0036ea3002e56b56226a561"} Dec 09 12:40:01 crc kubenswrapper[4770]: I1209 12:40:01.157646 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerStarted","Data":"bac4f791a7d2f241ddb92eab13d19ab8859a15e0b3bf13d8971016a47c74e63c"} Dec 09 12:40:01 crc kubenswrapper[4770]: I1209 12:40:01.184462 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-57mzf" podStartSLOduration=2.778587338 podStartE2EDuration="6.184440063s" podCreationTimestamp="2025-12-09 12:39:55 +0000 UTC" firstStartedPulling="2025-12-09 12:39:57.122842103 +0000 UTC m=+4122.363600622" lastFinishedPulling="2025-12-09 12:40:00.528694828 +0000 UTC m=+4125.769453347" observedRunningTime="2025-12-09 12:40:01.18038811 +0000 UTC m=+4126.421146639" watchObservedRunningTime="2025-12-09 12:40:01.184440063 +0000 UTC m=+4126.425198612" Dec 09 12:40:06 crc kubenswrapper[4770]: I1209 12:40:06.370563 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:06 crc kubenswrapper[4770]: I1209 12:40:06.372280 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:06 crc kubenswrapper[4770]: I1209 12:40:06.414996 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:07 crc kubenswrapper[4770]: I1209 12:40:07.241055 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:07 crc kubenswrapper[4770]: I1209 12:40:07.284857 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:40:09 crc kubenswrapper[4770]: I1209 12:40:09.217553 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-57mzf" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="registry-server" containerID="cri-o://bac4f791a7d2f241ddb92eab13d19ab8859a15e0b3bf13d8971016a47c74e63c" gracePeriod=2 Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.242629 4770 generic.go:334] "Generic (PLEG): container finished" podID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerID="bac4f791a7d2f241ddb92eab13d19ab8859a15e0b3bf13d8971016a47c74e63c" exitCode=0 Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.242703 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerDied","Data":"bac4f791a7d2f241ddb92eab13d19ab8859a15e0b3bf13d8971016a47c74e63c"} Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.243145 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57mzf" event={"ID":"92505208-1a8c-4b49-a0b4-49a8556f1e00","Type":"ContainerDied","Data":"4ab4aac0fdad7a15e99f495d5cf2cec15ea9f4a835502e63603356bd0b951aa8"} Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.243201 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ab4aac0fdad7a15e99f495d5cf2cec15ea9f4a835502e63603356bd0b951aa8" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.298235 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.446977 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities\") pod \"92505208-1a8c-4b49-a0b4-49a8556f1e00\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.447369 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content\") pod \"92505208-1a8c-4b49-a0b4-49a8556f1e00\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.447542 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntknq\" (UniqueName: \"kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq\") pod \"92505208-1a8c-4b49-a0b4-49a8556f1e00\" (UID: \"92505208-1a8c-4b49-a0b4-49a8556f1e00\") " Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.448055 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities" (OuterVolumeSpecName: "utilities") pod "92505208-1a8c-4b49-a0b4-49a8556f1e00" (UID: "92505208-1a8c-4b49-a0b4-49a8556f1e00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.455044 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq" (OuterVolumeSpecName: "kube-api-access-ntknq") pod "92505208-1a8c-4b49-a0b4-49a8556f1e00" (UID: "92505208-1a8c-4b49-a0b4-49a8556f1e00"). InnerVolumeSpecName "kube-api-access-ntknq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.549823 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.549875 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntknq\" (UniqueName: \"kubernetes.io/projected/92505208-1a8c-4b49-a0b4-49a8556f1e00-kube-api-access-ntknq\") on node \"crc\" DevicePath \"\"" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.562949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92505208-1a8c-4b49-a0b4-49a8556f1e00" (UID: "92505208-1a8c-4b49-a0b4-49a8556f1e00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:40:12 crc kubenswrapper[4770]: I1209 12:40:12.651670 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92505208-1a8c-4b49-a0b4-49a8556f1e00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:40:13 crc kubenswrapper[4770]: I1209 12:40:13.251291 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57mzf" Dec 09 12:40:13 crc kubenswrapper[4770]: I1209 12:40:13.288169 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:40:13 crc kubenswrapper[4770]: I1209 12:40:13.294473 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-57mzf"] Dec 09 12:40:13 crc kubenswrapper[4770]: I1209 12:40:13.350725 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" path="/var/lib/kubelet/pods/92505208-1a8c-4b49-a0b4-49a8556f1e00/volumes" Dec 09 12:41:02 crc kubenswrapper[4770]: I1209 12:41:02.474054 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:41:02 crc kubenswrapper[4770]: I1209 12:41:02.474649 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:41:32 crc kubenswrapper[4770]: I1209 12:41:32.473821 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:41:32 crc kubenswrapper[4770]: I1209 12:41:32.474425 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:42:02 crc kubenswrapper[4770]: I1209 12:42:02.473436 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:42:02 crc kubenswrapper[4770]: I1209 12:42:02.474037 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:42:02 crc kubenswrapper[4770]: I1209 12:42:02.474088 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:42:02 crc kubenswrapper[4770]: I1209 12:42:02.474695 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:42:02 crc kubenswrapper[4770]: I1209 12:42:02.474784 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d" gracePeriod=600 Dec 09 12:42:03 crc kubenswrapper[4770]: I1209 12:42:03.345774 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d" exitCode=0 Dec 09 12:42:03 crc kubenswrapper[4770]: I1209 12:42:03.350250 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d"} Dec 09 12:42:03 crc kubenswrapper[4770]: I1209 12:42:03.350290 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f"} Dec 09 12:42:03 crc kubenswrapper[4770]: I1209 12:42:03.350338 4770 scope.go:117] "RemoveContainer" containerID="f76b8fdb8339bf3c2237e7b53d48371169f1d09806b9b31a3c608d62cdb823fd" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.939595 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:06 crc kubenswrapper[4770]: E1209 12:42:06.940131 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="registry-server" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.940153 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="registry-server" Dec 09 12:42:06 crc kubenswrapper[4770]: E1209 12:42:06.940184 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="extract-utilities" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.940190 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="extract-utilities" Dec 09 12:42:06 crc kubenswrapper[4770]: E1209 12:42:06.940206 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="extract-content" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.940212 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="extract-content" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.940382 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="92505208-1a8c-4b49-a0b4-49a8556f1e00" containerName="registry-server" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.941620 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:06 crc kubenswrapper[4770]: I1209 12:42:06.952198 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.133100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4rlk\" (UniqueName: \"kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.133164 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.133245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.234610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.234706 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4rlk\" (UniqueName: \"kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.234724 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.235196 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.235233 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.258945 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4rlk\" (UniqueName: \"kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk\") pod \"community-operators-l9l52\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:07 crc kubenswrapper[4770]: I1209 12:42:07.345319 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:08 crc kubenswrapper[4770]: I1209 12:42:08.259164 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:08 crc kubenswrapper[4770]: I1209 12:42:08.383019 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerStarted","Data":"e916d24676883975a13f43e023a955059033fa893ae3633d17a4a2347348e863"} Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.391497 4770 generic.go:334] "Generic (PLEG): container finished" podID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerID="252400d1f24069bfc60e11f5710af34fc98c1456b477c6c44c064db62ec07a5a" exitCode=0 Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.391554 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerDied","Data":"252400d1f24069bfc60e11f5710af34fc98c1456b477c6c44c064db62ec07a5a"} Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.549981 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.552175 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.581118 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.689109 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzhjr\" (UniqueName: \"kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.689164 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.689183 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.791271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzhjr\" (UniqueName: \"kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.791339 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.791372 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.792106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.792162 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.815654 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzhjr\" (UniqueName: \"kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr\") pod \"redhat-marketplace-68ksb\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:09 crc kubenswrapper[4770]: I1209 12:42:09.886439 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:10 crc kubenswrapper[4770]: I1209 12:42:10.331357 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:10 crc kubenswrapper[4770]: I1209 12:42:10.399839 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerStarted","Data":"9f58b8daf7bf97f4cd3566cf4520768f9fa63043c885e4373e05a8b3601dc1cc"} Dec 09 12:42:10 crc kubenswrapper[4770]: I1209 12:42:10.402305 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerStarted","Data":"b05175d61f22353786e51e7b08d6d1309e544628b01e6bf19327efad1d451f45"} Dec 09 12:42:11 crc kubenswrapper[4770]: I1209 12:42:11.412071 4770 generic.go:334] "Generic (PLEG): container finished" podID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerID="61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635" exitCode=0 Dec 09 12:42:11 crc kubenswrapper[4770]: I1209 12:42:11.412125 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerDied","Data":"61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635"} Dec 09 12:42:11 crc kubenswrapper[4770]: I1209 12:42:11.415100 4770 generic.go:334] "Generic (PLEG): container finished" podID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerID="b05175d61f22353786e51e7b08d6d1309e544628b01e6bf19327efad1d451f45" exitCode=0 Dec 09 12:42:11 crc kubenswrapper[4770]: I1209 12:42:11.415155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerDied","Data":"b05175d61f22353786e51e7b08d6d1309e544628b01e6bf19327efad1d451f45"} Dec 09 12:42:12 crc kubenswrapper[4770]: I1209 12:42:12.478765 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerStarted","Data":"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b"} Dec 09 12:42:12 crc kubenswrapper[4770]: I1209 12:42:12.482887 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerStarted","Data":"7955bd92465385e8d13357b8dd451934b19c2c729b9f0abff1c27227604204d8"} Dec 09 12:42:13 crc kubenswrapper[4770]: I1209 12:42:13.493598 4770 generic.go:334] "Generic (PLEG): container finished" podID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerID="fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b" exitCode=0 Dec 09 12:42:13 crc kubenswrapper[4770]: I1209 12:42:13.493683 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerDied","Data":"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b"} Dec 09 12:42:13 crc kubenswrapper[4770]: I1209 12:42:13.523194 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l9l52" podStartSLOduration=4.973648654 podStartE2EDuration="7.523172285s" podCreationTimestamp="2025-12-09 12:42:06 +0000 UTC" firstStartedPulling="2025-12-09 12:42:09.393299377 +0000 UTC m=+4254.634057896" lastFinishedPulling="2025-12-09 12:42:11.942823008 +0000 UTC m=+4257.183581527" observedRunningTime="2025-12-09 12:42:12.523358734 +0000 UTC m=+4257.764117253" watchObservedRunningTime="2025-12-09 12:42:13.523172285 +0000 UTC m=+4258.763930804" Dec 09 12:42:14 crc kubenswrapper[4770]: I1209 12:42:14.527025 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerStarted","Data":"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185"} Dec 09 12:42:17 crc kubenswrapper[4770]: I1209 12:42:17.348719 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:17 crc kubenswrapper[4770]: I1209 12:42:17.348774 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:17 crc kubenswrapper[4770]: I1209 12:42:17.395967 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:17 crc kubenswrapper[4770]: I1209 12:42:17.420446 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-68ksb" podStartSLOduration=5.924970223 podStartE2EDuration="8.420423413s" podCreationTimestamp="2025-12-09 12:42:09 +0000 UTC" firstStartedPulling="2025-12-09 12:42:11.415703597 +0000 UTC m=+4256.656462116" lastFinishedPulling="2025-12-09 12:42:13.911156787 +0000 UTC m=+4259.151915306" observedRunningTime="2025-12-09 12:42:14.545039806 +0000 UTC m=+4259.785798325" watchObservedRunningTime="2025-12-09 12:42:17.420423413 +0000 UTC m=+4262.661181932" Dec 09 12:42:17 crc kubenswrapper[4770]: I1209 12:42:17.596936 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:18 crc kubenswrapper[4770]: I1209 12:42:18.329271 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:19 crc kubenswrapper[4770]: I1209 12:42:19.563478 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l9l52" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="registry-server" containerID="cri-o://7955bd92465385e8d13357b8dd451934b19c2c729b9f0abff1c27227604204d8" gracePeriod=2 Dec 09 12:42:19 crc kubenswrapper[4770]: I1209 12:42:19.887657 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:19 crc kubenswrapper[4770]: I1209 12:42:19.888318 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:19 crc kubenswrapper[4770]: I1209 12:42:19.941763 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:20 crc kubenswrapper[4770]: I1209 12:42:20.578977 4770 generic.go:334] "Generic (PLEG): container finished" podID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerID="7955bd92465385e8d13357b8dd451934b19c2c729b9f0abff1c27227604204d8" exitCode=0 Dec 09 12:42:20 crc kubenswrapper[4770]: I1209 12:42:20.579021 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerDied","Data":"7955bd92465385e8d13357b8dd451934b19c2c729b9f0abff1c27227604204d8"} Dec 09 12:42:20 crc kubenswrapper[4770]: I1209 12:42:20.626085 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.307986 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.386453 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content\") pod \"82d09aa2-f512-442c-90ed-652a372b2bf6\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.386592 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4rlk\" (UniqueName: \"kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk\") pod \"82d09aa2-f512-442c-90ed-652a372b2bf6\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.386712 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities\") pod \"82d09aa2-f512-442c-90ed-652a372b2bf6\" (UID: \"82d09aa2-f512-442c-90ed-652a372b2bf6\") " Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.387430 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities" (OuterVolumeSpecName: "utilities") pod "82d09aa2-f512-442c-90ed-652a372b2bf6" (UID: "82d09aa2-f512-442c-90ed-652a372b2bf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.391867 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk" (OuterVolumeSpecName: "kube-api-access-f4rlk") pod "82d09aa2-f512-442c-90ed-652a372b2bf6" (UID: "82d09aa2-f512-442c-90ed-652a372b2bf6"). InnerVolumeSpecName "kube-api-access-f4rlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.441713 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82d09aa2-f512-442c-90ed-652a372b2bf6" (UID: "82d09aa2-f512-442c-90ed-652a372b2bf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.488597 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.488644 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4rlk\" (UniqueName: \"kubernetes.io/projected/82d09aa2-f512-442c-90ed-652a372b2bf6-kube-api-access-f4rlk\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.488660 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82d09aa2-f512-442c-90ed-652a372b2bf6-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.591506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9l52" event={"ID":"82d09aa2-f512-442c-90ed-652a372b2bf6","Type":"ContainerDied","Data":"e916d24676883975a13f43e023a955059033fa893ae3633d17a4a2347348e863"} Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.591590 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9l52" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.591618 4770 scope.go:117] "RemoveContainer" containerID="7955bd92465385e8d13357b8dd451934b19c2c729b9f0abff1c27227604204d8" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.621766 4770 scope.go:117] "RemoveContainer" containerID="b05175d61f22353786e51e7b08d6d1309e544628b01e6bf19327efad1d451f45" Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.636933 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.643731 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l9l52"] Dec 09 12:42:21 crc kubenswrapper[4770]: I1209 12:42:21.653892 4770 scope.go:117] "RemoveContainer" containerID="252400d1f24069bfc60e11f5710af34fc98c1456b477c6c44c064db62ec07a5a" Dec 09 12:42:22 crc kubenswrapper[4770]: I1209 12:42:22.324644 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:22 crc kubenswrapper[4770]: I1209 12:42:22.600711 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-68ksb" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="registry-server" containerID="cri-o://e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185" gracePeriod=2 Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.191335 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.248810 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzhjr\" (UniqueName: \"kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr\") pod \"c30d9842-6546-4b4a-8b22-0cee6a94032e\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.248887 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities\") pod \"c30d9842-6546-4b4a-8b22-0cee6a94032e\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.248948 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content\") pod \"c30d9842-6546-4b4a-8b22-0cee6a94032e\" (UID: \"c30d9842-6546-4b4a-8b22-0cee6a94032e\") " Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.249864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities" (OuterVolumeSpecName: "utilities") pod "c30d9842-6546-4b4a-8b22-0cee6a94032e" (UID: "c30d9842-6546-4b4a-8b22-0cee6a94032e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.255300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr" (OuterVolumeSpecName: "kube-api-access-pzhjr") pod "c30d9842-6546-4b4a-8b22-0cee6a94032e" (UID: "c30d9842-6546-4b4a-8b22-0cee6a94032e"). InnerVolumeSpecName "kube-api-access-pzhjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.272449 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c30d9842-6546-4b4a-8b22-0cee6a94032e" (UID: "c30d9842-6546-4b4a-8b22-0cee6a94032e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.350814 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzhjr\" (UniqueName: \"kubernetes.io/projected/c30d9842-6546-4b4a-8b22-0cee6a94032e-kube-api-access-pzhjr\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.350852 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.350861 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30d9842-6546-4b4a-8b22-0cee6a94032e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.352115 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" path="/var/lib/kubelet/pods/82d09aa2-f512-442c-90ed-652a372b2bf6/volumes" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.609008 4770 generic.go:334] "Generic (PLEG): container finished" podID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerID="e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185" exitCode=0 Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.609056 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerDied","Data":"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185"} Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.609060 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68ksb" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.609103 4770 scope.go:117] "RemoveContainer" containerID="e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.609089 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68ksb" event={"ID":"c30d9842-6546-4b4a-8b22-0cee6a94032e","Type":"ContainerDied","Data":"9f58b8daf7bf97f4cd3566cf4520768f9fa63043c885e4373e05a8b3601dc1cc"} Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.638994 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.642298 4770 scope.go:117] "RemoveContainer" containerID="fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.646324 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-68ksb"] Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.657560 4770 scope.go:117] "RemoveContainer" containerID="61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.688155 4770 scope.go:117] "RemoveContainer" containerID="e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185" Dec 09 12:42:23 crc kubenswrapper[4770]: E1209 12:42:23.688659 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185\": container with ID starting with e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185 not found: ID does not exist" containerID="e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.688730 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185"} err="failed to get container status \"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185\": rpc error: code = NotFound desc = could not find container \"e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185\": container with ID starting with e55b7abb28f56648c7e05257dac466e647691371224f7a6999656a0b0b89e185 not found: ID does not exist" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.688770 4770 scope.go:117] "RemoveContainer" containerID="fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b" Dec 09 12:42:23 crc kubenswrapper[4770]: E1209 12:42:23.689638 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b\": container with ID starting with fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b not found: ID does not exist" containerID="fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.689696 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b"} err="failed to get container status \"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b\": rpc error: code = NotFound desc = could not find container \"fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b\": container with ID starting with fe1833e24aef30a101a8fb9e1b02591aa9b170493006a7cb21da6dd0777b2d5b not found: ID does not exist" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.689728 4770 scope.go:117] "RemoveContainer" containerID="61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635" Dec 09 12:42:23 crc kubenswrapper[4770]: E1209 12:42:23.690232 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635\": container with ID starting with 61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635 not found: ID does not exist" containerID="61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635" Dec 09 12:42:23 crc kubenswrapper[4770]: I1209 12:42:23.690272 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635"} err="failed to get container status \"61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635\": rpc error: code = NotFound desc = could not find container \"61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635\": container with ID starting with 61c1b1f2021fcc62f52edc6cd5401efdb5fd5117f8a1e2943370702668120635 not found: ID does not exist" Dec 09 12:42:25 crc kubenswrapper[4770]: I1209 12:42:25.350151 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" path="/var/lib/kubelet/pods/c30d9842-6546-4b4a-8b22-0cee6a94032e/volumes" Dec 09 12:44:02 crc kubenswrapper[4770]: I1209 12:44:02.474500 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:44:02 crc kubenswrapper[4770]: I1209 12:44:02.475170 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:44:32 crc kubenswrapper[4770]: I1209 12:44:32.474071 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:44:32 crc kubenswrapper[4770]: I1209 12:44:32.476090 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.211186 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86"] Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212358 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212394 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212423 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212430 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212439 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="extract-content" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212445 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="extract-content" Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212459 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="extract-utilities" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212465 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="extract-utilities" Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212474 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="extract-utilities" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212479 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="extract-utilities" Dec 09 12:45:00 crc kubenswrapper[4770]: E1209 12:45:00.212493 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="extract-content" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212499 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="extract-content" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212667 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d09aa2-f512-442c-90ed-652a372b2bf6" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.212686 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c30d9842-6546-4b4a-8b22-0cee6a94032e" containerName="registry-server" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.215921 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.218673 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.218707 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.223516 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86"] Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.282395 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.282555 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sqfv\" (UniqueName: \"kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.282776 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.384633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.384710 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.384769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sqfv\" (UniqueName: \"kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.386145 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.401257 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.411713 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sqfv\" (UniqueName: \"kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv\") pod \"collect-profiles-29421405-kdc86\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.547586 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:00 crc kubenswrapper[4770]: I1209 12:45:00.993163 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86"] Dec 09 12:45:01 crc kubenswrapper[4770]: I1209 12:45:01.102070 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" event={"ID":"5c87afe1-1316-4bf9-87a6-8e479c43c0b6","Type":"ContainerStarted","Data":"3d4d3790aa12882436355eeb1f39c72cfda71dbd0f8010985f4cb2ef1c8e2b0e"} Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.111684 4770 generic.go:334] "Generic (PLEG): container finished" podID="5c87afe1-1316-4bf9-87a6-8e479c43c0b6" containerID="6f3b6b2bdf5ef33c728bf3f80ea1e0d27470fb51fa639d24c6307c9434f8c352" exitCode=0 Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.111796 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" event={"ID":"5c87afe1-1316-4bf9-87a6-8e479c43c0b6","Type":"ContainerDied","Data":"6f3b6b2bdf5ef33c728bf3f80ea1e0d27470fb51fa639d24c6307c9434f8c352"} Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.473845 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.473928 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.473991 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.474748 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:45:02 crc kubenswrapper[4770]: I1209 12:45:02.474829 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" gracePeriod=600 Dec 09 12:45:02 crc kubenswrapper[4770]: E1209 12:45:02.596694 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.123009 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" exitCode=0 Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.123052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f"} Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.123163 4770 scope.go:117] "RemoveContainer" containerID="405156e42128824348e13c0378ff59f43c0e909a78a7894930b2fb24ccc2286d" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.123882 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:45:03 crc kubenswrapper[4770]: E1209 12:45:03.124148 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.579388 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.623690 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume\") pod \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.623827 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume\") pod \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.623859 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sqfv\" (UniqueName: \"kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv\") pod \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\" (UID: \"5c87afe1-1316-4bf9-87a6-8e479c43c0b6\") " Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.624648 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5c87afe1-1316-4bf9-87a6-8e479c43c0b6" (UID: "5c87afe1-1316-4bf9-87a6-8e479c43c0b6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.629666 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv" (OuterVolumeSpecName: "kube-api-access-5sqfv") pod "5c87afe1-1316-4bf9-87a6-8e479c43c0b6" (UID: "5c87afe1-1316-4bf9-87a6-8e479c43c0b6"). InnerVolumeSpecName "kube-api-access-5sqfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.630015 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5c87afe1-1316-4bf9-87a6-8e479c43c0b6" (UID: "5c87afe1-1316-4bf9-87a6-8e479c43c0b6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.725549 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.725599 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 12:45:03 crc kubenswrapper[4770]: I1209 12:45:03.725614 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sqfv\" (UniqueName: \"kubernetes.io/projected/5c87afe1-1316-4bf9-87a6-8e479c43c0b6-kube-api-access-5sqfv\") on node \"crc\" DevicePath \"\"" Dec 09 12:45:04 crc kubenswrapper[4770]: I1209 12:45:04.133148 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" event={"ID":"5c87afe1-1316-4bf9-87a6-8e479c43c0b6","Type":"ContainerDied","Data":"3d4d3790aa12882436355eeb1f39c72cfda71dbd0f8010985f4cb2ef1c8e2b0e"} Dec 09 12:45:04 crc kubenswrapper[4770]: I1209 12:45:04.133213 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421405-kdc86" Dec 09 12:45:04 crc kubenswrapper[4770]: I1209 12:45:04.133212 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4d3790aa12882436355eeb1f39c72cfda71dbd0f8010985f4cb2ef1c8e2b0e" Dec 09 12:45:04 crc kubenswrapper[4770]: I1209 12:45:04.688075 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm"] Dec 09 12:45:04 crc kubenswrapper[4770]: I1209 12:45:04.694698 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421360-m2gjm"] Dec 09 12:45:05 crc kubenswrapper[4770]: I1209 12:45:05.350430 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df1e7e38-0196-43f1-a5f3-b149ef929584" path="/var/lib/kubelet/pods/df1e7e38-0196-43f1-a5f3-b149ef929584/volumes" Dec 09 12:45:17 crc kubenswrapper[4770]: I1209 12:45:17.340553 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:45:17 crc kubenswrapper[4770]: E1209 12:45:17.341290 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:45:29 crc kubenswrapper[4770]: I1209 12:45:29.340150 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:45:29 crc kubenswrapper[4770]: E1209 12:45:29.341008 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:45:31 crc kubenswrapper[4770]: I1209 12:45:31.587813 4770 scope.go:117] "RemoveContainer" containerID="b7393b5b593402a6cad01634e574580bc6022d654143620f6f64057329e39af6" Dec 09 12:45:42 crc kubenswrapper[4770]: I1209 12:45:42.340629 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:45:42 crc kubenswrapper[4770]: E1209 12:45:42.344137 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:45:56 crc kubenswrapper[4770]: I1209 12:45:56.345008 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:45:56 crc kubenswrapper[4770]: E1209 12:45:56.346142 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:46:09 crc kubenswrapper[4770]: I1209 12:46:09.340435 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:46:09 crc kubenswrapper[4770]: E1209 12:46:09.341392 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:46:22 crc kubenswrapper[4770]: I1209 12:46:22.340497 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:46:22 crc kubenswrapper[4770]: E1209 12:46:22.341189 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:46:31 crc kubenswrapper[4770]: I1209 12:46:31.645060 4770 scope.go:117] "RemoveContainer" containerID="d0628aee28c653c06a67e4d0a81a53c216405d3cf0036ea3002e56b56226a561" Dec 09 12:46:31 crc kubenswrapper[4770]: I1209 12:46:31.677534 4770 scope.go:117] "RemoveContainer" containerID="bac4f791a7d2f241ddb92eab13d19ab8859a15e0b3bf13d8971016a47c74e63c" Dec 09 12:46:31 crc kubenswrapper[4770]: I1209 12:46:31.711375 4770 scope.go:117] "RemoveContainer" containerID="30a2fedba9430ce527d49f9bc9cc9679f048495e1c649f68e3113b451961e28b" Dec 09 12:46:37 crc kubenswrapper[4770]: I1209 12:46:37.340434 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:46:37 crc kubenswrapper[4770]: E1209 12:46:37.341476 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:46:52 crc kubenswrapper[4770]: I1209 12:46:52.398860 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:46:52 crc kubenswrapper[4770]: E1209 12:46:52.399924 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:47:04 crc kubenswrapper[4770]: I1209 12:47:04.340391 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:47:04 crc kubenswrapper[4770]: E1209 12:47:04.341111 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:47:19 crc kubenswrapper[4770]: I1209 12:47:19.341704 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:47:19 crc kubenswrapper[4770]: E1209 12:47:19.343742 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:47:31 crc kubenswrapper[4770]: I1209 12:47:31.340881 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:47:31 crc kubenswrapper[4770]: E1209 12:47:31.342193 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:47:42 crc kubenswrapper[4770]: I1209 12:47:42.340219 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:47:42 crc kubenswrapper[4770]: E1209 12:47:42.340878 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:47:55 crc kubenswrapper[4770]: I1209 12:47:55.346065 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:47:55 crc kubenswrapper[4770]: E1209 12:47:55.346955 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:48:07 crc kubenswrapper[4770]: I1209 12:48:07.340291 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:48:07 crc kubenswrapper[4770]: E1209 12:48:07.341069 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:48:21 crc kubenswrapper[4770]: I1209 12:48:21.341126 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:48:21 crc kubenswrapper[4770]: E1209 12:48:21.342000 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:48:32 crc kubenswrapper[4770]: I1209 12:48:32.340184 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:48:32 crc kubenswrapper[4770]: E1209 12:48:32.340892 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:48:47 crc kubenswrapper[4770]: I1209 12:48:47.340633 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:48:47 crc kubenswrapper[4770]: E1209 12:48:47.341570 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:02 crc kubenswrapper[4770]: I1209 12:49:02.340363 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:49:02 crc kubenswrapper[4770]: E1209 12:49:02.341434 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:15 crc kubenswrapper[4770]: I1209 12:49:15.348991 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:49:15 crc kubenswrapper[4770]: E1209 12:49:15.349719 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:29 crc kubenswrapper[4770]: I1209 12:49:29.341646 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:49:29 crc kubenswrapper[4770]: E1209 12:49:29.342846 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:44 crc kubenswrapper[4770]: I1209 12:49:44.339921 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:49:44 crc kubenswrapper[4770]: E1209 12:49:44.340635 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:56 crc kubenswrapper[4770]: I1209 12:49:56.340709 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:49:56 crc kubenswrapper[4770]: E1209 12:49:56.346863 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.880037 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2pcrp/must-gather-cxm8p"] Dec 09 12:49:57 crc kubenswrapper[4770]: E1209 12:49:57.880625 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c87afe1-1316-4bf9-87a6-8e479c43c0b6" containerName="collect-profiles" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.880646 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c87afe1-1316-4bf9-87a6-8e479c43c0b6" containerName="collect-profiles" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.880832 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c87afe1-1316-4bf9-87a6-8e479c43c0b6" containerName="collect-profiles" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.881668 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.883151 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2pcrp"/"default-dockercfg-hqldk" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.883930 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2pcrp"/"openshift-service-ca.crt" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.884194 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2pcrp"/"kube-root-ca.crt" Dec 09 12:49:57 crc kubenswrapper[4770]: I1209 12:49:57.900513 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2pcrp/must-gather-cxm8p"] Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.030544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.030601 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcfkk\" (UniqueName: \"kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.132373 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.132425 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcfkk\" (UniqueName: \"kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.133171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.154772 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcfkk\" (UniqueName: \"kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk\") pod \"must-gather-cxm8p\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:58 crc kubenswrapper[4770]: I1209 12:49:58.203383 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:49:59 crc kubenswrapper[4770]: I1209 12:49:59.075140 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2pcrp/must-gather-cxm8p"] Dec 09 12:49:59 crc kubenswrapper[4770]: I1209 12:49:59.087745 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 12:50:00 crc kubenswrapper[4770]: I1209 12:50:00.071983 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" event={"ID":"c1467a75-7b8e-4b1d-a56e-9529a26a1c08","Type":"ContainerStarted","Data":"0564fc52831798598729351da55c32f0a874ba453141657ea535193a04359023"} Dec 09 12:50:06 crc kubenswrapper[4770]: I1209 12:50:06.122024 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" event={"ID":"c1467a75-7b8e-4b1d-a56e-9529a26a1c08","Type":"ContainerStarted","Data":"fd2abc20aca6fe97ef4c9ec3962f49c4c5ab1afc25959ecdb9a7ee4f114ab9b6"} Dec 09 12:50:06 crc kubenswrapper[4770]: I1209 12:50:06.122621 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" event={"ID":"c1467a75-7b8e-4b1d-a56e-9529a26a1c08","Type":"ContainerStarted","Data":"17a9922e58d554d52e910815c39297ca4ee464ebd94048d073bc85a3c132abb3"} Dec 09 12:50:06 crc kubenswrapper[4770]: I1209 12:50:06.137169 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" podStartSLOduration=2.547517821 podStartE2EDuration="9.13714072s" podCreationTimestamp="2025-12-09 12:49:57 +0000 UTC" firstStartedPulling="2025-12-09 12:49:59.087452614 +0000 UTC m=+4724.328211133" lastFinishedPulling="2025-12-09 12:50:05.677075513 +0000 UTC m=+4730.917834032" observedRunningTime="2025-12-09 12:50:06.136040712 +0000 UTC m=+4731.376799231" watchObservedRunningTime="2025-12-09 12:50:06.13714072 +0000 UTC m=+4731.377899239" Dec 09 12:50:07 crc kubenswrapper[4770]: I1209 12:50:07.340108 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:50:08 crc kubenswrapper[4770]: I1209 12:50:08.148333 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6"} Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.251444 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.255271 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.279323 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.526511 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqkkd\" (UniqueName: \"kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.527135 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.527163 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.628171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.628220 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.628305 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqkkd\" (UniqueName: \"kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.628756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.628993 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.668102 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqkkd\" (UniqueName: \"kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd\") pod \"redhat-operators-dgkvp\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:09 crc kubenswrapper[4770]: I1209 12:50:09.879015 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:10 crc kubenswrapper[4770]: I1209 12:50:10.523046 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:10 crc kubenswrapper[4770]: W1209 12:50:10.532098 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d1d3ae_1167_4af2_85fe_c2931e8ab8f8.slice/crio-0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2 WatchSource:0}: Error finding container 0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2: Status 404 returned error can't find the container with id 0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2 Dec 09 12:50:11 crc kubenswrapper[4770]: I1209 12:50:11.204102 4770 generic.go:334] "Generic (PLEG): container finished" podID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerID="4263ac399530190891244e9dd73ee6d2b37cb0d0238032c1a6ff33a4421646b9" exitCode=0 Dec 09 12:50:11 crc kubenswrapper[4770]: I1209 12:50:11.204200 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerDied","Data":"4263ac399530190891244e9dd73ee6d2b37cb0d0238032c1a6ff33a4421646b9"} Dec 09 12:50:11 crc kubenswrapper[4770]: I1209 12:50:11.204742 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerStarted","Data":"0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2"} Dec 09 12:50:12 crc kubenswrapper[4770]: I1209 12:50:12.219722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerStarted","Data":"350e709db9ee32fbb6a5dabcc17ed5a2d9aed3e20d3df253cdf872c879aee0c3"} Dec 09 12:50:14 crc kubenswrapper[4770]: I1209 12:50:14.237194 4770 generic.go:334] "Generic (PLEG): container finished" podID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerID="350e709db9ee32fbb6a5dabcc17ed5a2d9aed3e20d3df253cdf872c879aee0c3" exitCode=0 Dec 09 12:50:14 crc kubenswrapper[4770]: I1209 12:50:14.237293 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerDied","Data":"350e709db9ee32fbb6a5dabcc17ed5a2d9aed3e20d3df253cdf872c879aee0c3"} Dec 09 12:50:15 crc kubenswrapper[4770]: I1209 12:50:15.248112 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerStarted","Data":"88031c6defa364721c70c5e92c7d32636a9b24a41cc24c03860cd5e1dc780bad"} Dec 09 12:50:15 crc kubenswrapper[4770]: I1209 12:50:15.276251 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dgkvp" podStartSLOduration=2.837140194 podStartE2EDuration="6.276229881s" podCreationTimestamp="2025-12-09 12:50:09 +0000 UTC" firstStartedPulling="2025-12-09 12:50:11.206499671 +0000 UTC m=+4736.447258180" lastFinishedPulling="2025-12-09 12:50:14.645589348 +0000 UTC m=+4739.886347867" observedRunningTime="2025-12-09 12:50:15.2694789 +0000 UTC m=+4740.510237429" watchObservedRunningTime="2025-12-09 12:50:15.276229881 +0000 UTC m=+4740.516988410" Dec 09 12:50:19 crc kubenswrapper[4770]: I1209 12:50:19.879730 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:19 crc kubenswrapper[4770]: I1209 12:50:19.880342 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:20 crc kubenswrapper[4770]: I1209 12:50:20.927878 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dgkvp" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="registry-server" probeResult="failure" output=< Dec 09 12:50:20 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Dec 09 12:50:20 crc kubenswrapper[4770]: > Dec 09 12:50:29 crc kubenswrapper[4770]: I1209 12:50:29.928152 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:29 crc kubenswrapper[4770]: I1209 12:50:29.978541 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:30 crc kubenswrapper[4770]: I1209 12:50:30.166788 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:31 crc kubenswrapper[4770]: I1209 12:50:31.504741 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dgkvp" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="registry-server" containerID="cri-o://88031c6defa364721c70c5e92c7d32636a9b24a41cc24c03860cd5e1dc780bad" gracePeriod=2 Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.515088 4770 generic.go:334] "Generic (PLEG): container finished" podID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerID="88031c6defa364721c70c5e92c7d32636a9b24a41cc24c03860cd5e1dc780bad" exitCode=0 Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.515136 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerDied","Data":"88031c6defa364721c70c5e92c7d32636a9b24a41cc24c03860cd5e1dc780bad"} Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.515404 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgkvp" event={"ID":"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8","Type":"ContainerDied","Data":"0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2"} Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.515433 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e757b592722367f518ad50ff61d390bacd51122e933e08bbd13dbbfb810b9a2" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.519703 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.697832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities\") pod \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.697997 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqkkd\" (UniqueName: \"kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd\") pod \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.698065 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content\") pod \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\" (UID: \"a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8\") " Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.698643 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities" (OuterVolumeSpecName: "utilities") pod "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" (UID: "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.715740 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd" (OuterVolumeSpecName: "kube-api-access-hqkkd") pod "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" (UID: "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8"). InnerVolumeSpecName "kube-api-access-hqkkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.799528 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.799589 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqkkd\" (UniqueName: \"kubernetes.io/projected/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-kube-api-access-hqkkd\") on node \"crc\" DevicePath \"\"" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.813445 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" (UID: "a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:50:32 crc kubenswrapper[4770]: I1209 12:50:32.901458 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:50:33 crc kubenswrapper[4770]: I1209 12:50:33.533416 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgkvp" Dec 09 12:50:33 crc kubenswrapper[4770]: I1209 12:50:33.559048 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:33 crc kubenswrapper[4770]: I1209 12:50:33.566691 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dgkvp"] Dec 09 12:50:35 crc kubenswrapper[4770]: I1209 12:50:35.358269 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" path="/var/lib/kubelet/pods/a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8/volumes" Dec 09 12:51:06 crc kubenswrapper[4770]: I1209 12:51:06.645969 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/util/0.log" Dec 09 12:51:06 crc kubenswrapper[4770]: I1209 12:51:06.810471 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/pull/0.log" Dec 09 12:51:06 crc kubenswrapper[4770]: I1209 12:51:06.816380 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/util/0.log" Dec 09 12:51:06 crc kubenswrapper[4770]: I1209 12:51:06.845194 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/pull/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.240800 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/pull/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.242602 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/util/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.243368 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3xf9qx_c7f8c22f-e13b-48d4-868a-3e2fbd2e6eaa/extract/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.382595 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h5w2l_303a0c89-f011-4953-ac5c-33929697ccd1/kube-rbac-proxy/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.413343 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-5gj7z_cf58af47-dce9-43df-90cc-7c642e3b2f3b/kube-rbac-proxy/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.455547 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-h5w2l_303a0c89-f011-4953-ac5c-33929697ccd1/manager/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.589688 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-r89wl_9a0fffd2-c599-4a15-b1f0-3c404d44a7bc/kube-rbac-proxy/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.598575 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-5gj7z_cf58af47-dce9-43df-90cc-7c642e3b2f3b/manager/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.622177 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-r89wl_9a0fffd2-c599-4a15-b1f0-3c404d44a7bc/manager/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.756081 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-p69k8_3588fb02-b477-40dd-976e-f9b8e0508324/kube-rbac-proxy/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.860463 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-p69k8_3588fb02-b477-40dd-976e-f9b8e0508324/manager/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.953793 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-56gd2_2537825c-aa47-40ac-bb98-d9793449e9dd/kube-rbac-proxy/0.log" Dec 09 12:51:07 crc kubenswrapper[4770]: I1209 12:51:07.962498 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-56gd2_2537825c-aa47-40ac-bb98-d9793449e9dd/manager/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.071918 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4lvc9_23265ff6-b252-4fc0-a3b4-6b59deea7c69/kube-rbac-proxy/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.339294 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-4lvc9_23265ff6-b252-4fc0-a3b4-6b59deea7c69/manager/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.418721 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-kf5v2_d241ba8f-5573-466a-88f2-89ec9f14cc29/kube-rbac-proxy/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.646203 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-kf5v2_d241ba8f-5573-466a-88f2-89ec9f14cc29/manager/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.670072 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-zhc7c_c8cbcbfa-fe34-4fae-96b6-04a413418bf0/manager/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.692191 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-zhc7c_c8cbcbfa-fe34-4fae-96b6-04a413418bf0/kube-rbac-proxy/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.860139 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-846ng_7a873c67-c1eb-4681-b18e-25e1580bb11c/kube-rbac-proxy/0.log" Dec 09 12:51:08 crc kubenswrapper[4770]: I1209 12:51:08.874748 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-846ng_7a873c67-c1eb-4681-b18e-25e1580bb11c/manager/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.001791 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vvf9z_211dd097-458c-435c-b3b0-b952dc866fac/kube-rbac-proxy/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.051602 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vvf9z_211dd097-458c-435c-b3b0-b952dc866fac/manager/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.071338 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-5jr9q_fd01836c-e495-43e0-8552-e17a41352a3d/kube-rbac-proxy/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.520376 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-5jr9q_fd01836c-e495-43e0-8552-e17a41352a3d/manager/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.556114 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-jv56l_8c0e510c-adb0-43f4-b11d-05b378876a46/kube-rbac-proxy/0.log" Dec 09 12:51:09 crc kubenswrapper[4770]: I1209 12:51:09.564066 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-jv56l_8c0e510c-adb0-43f4-b11d-05b378876a46/manager/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.359802 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vjw9d_520a9036-0b49-4236-90cf-ada8df687ad9/kube-rbac-proxy/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.468446 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wljmn_7eea33f2-55bf-4542-82db-7cf987f29a63/kube-rbac-proxy/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.475306 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vjw9d_520a9036-0b49-4236-90cf-ada8df687ad9/manager/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.699525 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-wljmn_7eea33f2-55bf-4542-82db-7cf987f29a63/manager/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.775045 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4_86d62982-f66a-4184-8e15-9336bfca2b8f/kube-rbac-proxy/0.log" Dec 09 12:51:11 crc kubenswrapper[4770]: I1209 12:51:11.777078 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-694d6cfbd6c68s4_86d62982-f66a-4184-8e15-9336bfca2b8f/manager/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.127136 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-c4v65_15ff9cb7-d6ab-43d9-bddf-4efa94d8ad7a/registry-server/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.184917 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7979d445b4-qvrn5_ddbcaf12-235d-4b39-ba48-35c3fb15b5a8/operator/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.263177 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-w86cd_764304ee-eb30-434a-b884-f7e455576411/kube-rbac-proxy/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.392391 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-w86cd_764304ee-eb30-434a-b884-f7e455576411/manager/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.439593 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xszh4_70ea2809-e1eb-4cf6-bd48-21ad18c031a4/kube-rbac-proxy/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.526414 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-xszh4_70ea2809-e1eb-4cf6-bd48-21ad18c031a4/manager/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.813990 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-h6w6h_30458aa5-ed86-491e-b2bd-353725caf57d/operator/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.817822 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-668858c49-vsvv8_c57a5114-2563-4390-855b-402e73c907d6/manager/0.log" Dec 09 12:51:12 crc kubenswrapper[4770]: I1209 12:51:12.898498 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-b7x22_3da14cb0-e4a9-4fe0-bf66-616a2ed438eb/kube-rbac-proxy/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.009580 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-b7x22_3da14cb0-e4a9-4fe0-bf66-616a2ed438eb/manager/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.029226 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-96cv7_daa9d1b4-0867-4e4b-a5a4-037de4506c62/kube-rbac-proxy/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.077160 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-96cv7_daa9d1b4-0867-4e4b-a5a4-037de4506c62/manager/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.176542 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bggpm_049491b6-ec1f-421a-bfb3-bc7c63f27f8c/kube-rbac-proxy/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.286784 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-bq5ts_7af1b779-e5fd-40a2-8913-9a334a300efb/kube-rbac-proxy/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.293854 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bggpm_049491b6-ec1f-421a-bfb3-bc7c63f27f8c/manager/0.log" Dec 09 12:51:13 crc kubenswrapper[4770]: I1209 12:51:13.339146 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-bq5ts_7af1b779-e5fd-40a2-8913-9a334a300efb/manager/0.log" Dec 09 12:51:25 crc kubenswrapper[4770]: I1209 12:51:25.534159 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 12:51:25 crc kubenswrapper[4770]: I1209 12:51:25.535260 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-b7x22" podUID="3da14cb0-e4a9-4fe0-bf66-616a2ed438eb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 12:51:33 crc kubenswrapper[4770]: I1209 12:51:33.402724 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bhktd_4e08edb8-6c28-49a2-ba80-7e3703f48f5f/control-plane-machine-set-operator/0.log" Dec 09 12:51:33 crc kubenswrapper[4770]: I1209 12:51:33.530885 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5nk5g_7ed604a1-90f7-4f46-a4a3-8e097b02c922/kube-rbac-proxy/0.log" Dec 09 12:51:33 crc kubenswrapper[4770]: I1209 12:51:33.595526 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5nk5g_7ed604a1-90f7-4f46-a4a3-8e097b02c922/machine-api-operator/0.log" Dec 09 12:51:46 crc kubenswrapper[4770]: I1209 12:51:46.279283 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-zxvr2_b28fbeaa-16cf-4e85-bc3d-318554a2c422/cert-manager-controller/0.log" Dec 09 12:51:46 crc kubenswrapper[4770]: I1209 12:51:46.463390 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-ksfw7_9ae62b97-709d-4d61-8067-808e46f66cc0/cert-manager-cainjector/0.log" Dec 09 12:51:46 crc kubenswrapper[4770]: I1209 12:51:46.576267 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-w75jb_07764a7a-7c6b-4839-a163-695ce4b71961/cert-manager-webhook/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.415083 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-84prv_e684363b-6fd1-4432-8cee-cdb9858e4367/nmstate-console-plugin/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.634248 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dpkk8_e48df172-3e28-47a9-acf7-80d69988cd6b/nmstate-handler/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.714453 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-8s7rs_86ad5bcc-54e9-4125-a7d7-db0c74de1b01/kube-rbac-proxy/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.793156 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-8s7rs_86ad5bcc-54e9-4125-a7d7-db0c74de1b01/nmstate-metrics/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.893670 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-twvbf_094c86c2-b555-44b7-b26f-6915f0a5eaa5/nmstate-operator/0.log" Dec 09 12:52:00 crc kubenswrapper[4770]: I1209 12:52:00.960610 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-6p6bz_ab4bae62-c176-4ff2-89af-2b9a362c4f65/nmstate-webhook/0.log" Dec 09 12:52:17 crc kubenswrapper[4770]: I1209 12:52:17.983443 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n9pcb_844a755c-0f78-40a6-93f3-a388dce59835/kube-rbac-proxy/0.log" Dec 09 12:52:18 crc kubenswrapper[4770]: I1209 12:52:18.348812 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n9pcb_844a755c-0f78-40a6-93f3-a388dce59835/controller/0.log" Dec 09 12:52:18 crc kubenswrapper[4770]: I1209 12:52:18.482757 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-frr-files/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.070773 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-metrics/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.106767 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-reloader/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.127001 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-reloader/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.127992 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-frr-files/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.367784 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-reloader/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.369381 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-metrics/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.405631 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-frr-files/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.451942 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-metrics/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.740993 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-frr-files/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.759843 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-metrics/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.771181 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/cp-reloader/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.783687 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/controller/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.956952 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/frr-metrics/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.978847 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/kube-rbac-proxy-frr/0.log" Dec 09 12:52:19 crc kubenswrapper[4770]: I1209 12:52:19.983957 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/kube-rbac-proxy/0.log" Dec 09 12:52:20 crc kubenswrapper[4770]: I1209 12:52:20.173549 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/reloader/0.log" Dec 09 12:52:20 crc kubenswrapper[4770]: I1209 12:52:20.290731 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-jlj84_4b906317-0543-467d-9e51-a3af3f40d6b6/frr-k8s-webhook-server/0.log" Dec 09 12:52:20 crc kubenswrapper[4770]: I1209 12:52:20.424570 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7b5dcdff77-vbqkc_503c0654-d161-4668-a6d5-aaca43468503/manager/0.log" Dec 09 12:52:20 crc kubenswrapper[4770]: I1209 12:52:20.654564 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7b9c5877dd-p87cb_13dad646-e3a3-40ea-9a0f-6892bfaae9cd/webhook-server/0.log" Dec 09 12:52:20 crc kubenswrapper[4770]: I1209 12:52:20.692966 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7kzks_e013d4c8-5587-407a-9a8f-b26ab7064ec8/kube-rbac-proxy/0.log" Dec 09 12:52:21 crc kubenswrapper[4770]: I1209 12:52:21.360912 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7kzks_e013d4c8-5587-407a-9a8f-b26ab7064ec8/speaker/0.log" Dec 09 12:52:21 crc kubenswrapper[4770]: I1209 12:52:21.452339 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8b799_4002944e-6d2f-4f2b-8d01-e664da3116e2/frr/0.log" Dec 09 12:52:32 crc kubenswrapper[4770]: I1209 12:52:32.473392 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:52:32 crc kubenswrapper[4770]: I1209 12:52:32.473945 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:52:34 crc kubenswrapper[4770]: I1209 12:52:34.991466 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/util/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.136874 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/util/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.233981 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/pull/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.242225 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/pull/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.384176 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/util/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.416408 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/extract/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.434403 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ar69x9_00e48fd8-562c-4212-82c4-40101961789b/pull/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.564594 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/util/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.746291 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/pull/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.753178 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/util/0.log" Dec 09 12:52:35 crc kubenswrapper[4770]: I1209 12:52:35.753611 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/pull/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.265537 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/pull/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.273959 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/util/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.313648 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg97k2_91ff00b8-1682-47e7-99d8-b0444921b1a1/extract/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.458181 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/util/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.639122 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/pull/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.659742 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/util/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.660500 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/pull/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.809920 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/util/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.820890 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/pull/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.822716 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83vcr9z_2b7353bb-7d5f-47e4-99c5-b319206fcb90/extract/0.log" Dec 09 12:52:36 crc kubenswrapper[4770]: I1209 12:52:36.971822 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-utilities/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.142583 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-content/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.149472 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-utilities/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.159011 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-content/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.313561 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-utilities/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.349477 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/extract-content/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.518569 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-utilities/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.736179 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-content/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.753465 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-content/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.799674 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-utilities/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.971068 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vn6tj_0111570a-aa10-4ffb-a876-9b24fcc9415a/registry-server/0.log" Dec 09 12:52:37 crc kubenswrapper[4770]: I1209 12:52:37.987238 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-utilities/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.041646 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/extract-content/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.167890 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-twm9p_9e029e5b-b9b0-483a-ad49-29f8ceb43e61/marketplace-operator/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.616472 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-utilities/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.737271 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-utilities/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.772436 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-content/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.835772 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2b6tg_ac16f7d6-2737-46df-8cf8-6df5fafbb9c6/registry-server/0.log" Dec 09 12:52:38 crc kubenswrapper[4770]: I1209 12:52:38.840394 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-content/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.013410 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-utilities/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.026992 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/extract-content/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.121103 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-utilities/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.297918 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qfsls_d42dfe0d-819e-48a0-b808-e2a34f531355/registry-server/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.324813 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-content/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.339687 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-utilities/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.365671 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-content/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.498601 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-utilities/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.512125 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/extract-content/0.log" Dec 09 12:52:39 crc kubenswrapper[4770]: I1209 12:52:39.836045 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7tr6q_0842ce6d-5d86-4008-ad0a-78674b554630/registry-server/0.log" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.900340 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:42 crc kubenswrapper[4770]: E1209 12:52:42.900715 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="registry-server" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.900733 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="registry-server" Dec 09 12:52:42 crc kubenswrapper[4770]: E1209 12:52:42.900768 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="extract-utilities" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.900775 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="extract-utilities" Dec 09 12:52:42 crc kubenswrapper[4770]: E1209 12:52:42.900798 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="extract-content" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.900804 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="extract-content" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.901010 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d1d3ae-1167-4af2-85fe-c2931e8ab8f8" containerName="registry-server" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.902320 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:42 crc kubenswrapper[4770]: I1209 12:52:42.914226 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.062894 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.063684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndkx4\" (UniqueName: \"kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.064076 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.100789 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.102641 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.118438 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.165072 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.165151 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.165183 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndkx4\" (UniqueName: \"kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.166151 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.166415 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.190051 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndkx4\" (UniqueName: \"kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4\") pod \"redhat-marketplace-2xm8j\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.226706 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.266602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258ll\" (UniqueName: \"kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.266706 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.266753 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.368532 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.368592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.368650 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258ll\" (UniqueName: \"kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.369422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.370179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.391975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258ll\" (UniqueName: \"kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll\") pod \"certified-operators-xwkhb\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.419318 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.549335 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:43 crc kubenswrapper[4770]: I1209 12:52:43.932845 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:43 crc kubenswrapper[4770]: W1209 12:52:43.935126 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod239014c3_d863_42ab_8f3d_62f19934afba.slice/crio-07a0f2af9fa97b27820ce78b0371b39268060c1a3d5ccd08ce4c9ec21db2a14e WatchSource:0}: Error finding container 07a0f2af9fa97b27820ce78b0371b39268060c1a3d5ccd08ce4c9ec21db2a14e: Status 404 returned error can't find the container with id 07a0f2af9fa97b27820ce78b0371b39268060c1a3d5ccd08ce4c9ec21db2a14e Dec 09 12:52:44 crc kubenswrapper[4770]: I1209 12:52:44.000937 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerStarted","Data":"07a0f2af9fa97b27820ce78b0371b39268060c1a3d5ccd08ce4c9ec21db2a14e"} Dec 09 12:52:44 crc kubenswrapper[4770]: I1209 12:52:44.002752 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc8c352a-3f30-4fc3-bb35-fe518cd4d940" containerID="fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca" exitCode=0 Dec 09 12:52:44 crc kubenswrapper[4770]: I1209 12:52:44.002817 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerDied","Data":"fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca"} Dec 09 12:52:44 crc kubenswrapper[4770]: I1209 12:52:44.002919 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerStarted","Data":"aecddccaab60048b49f546c6fb79d05210bcb70428cc323b1b5b1569dffd3aec"} Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.021797 4770 generic.go:334] "Generic (PLEG): container finished" podID="239014c3-d863-42ab-8f3d-62f19934afba" containerID="a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c" exitCode=0 Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.021896 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerDied","Data":"a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c"} Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.025218 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc8c352a-3f30-4fc3-bb35-fe518cd4d940" containerID="76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8" exitCode=0 Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.025259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerDied","Data":"76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8"} Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.299250 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.300853 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.317828 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.398988 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.399055 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kp98\" (UniqueName: \"kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.399368 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.500267 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.500313 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kp98\" (UniqueName: \"kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.500379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.500911 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.500886 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.527279 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kp98\" (UniqueName: \"kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98\") pod \"community-operators-kpk4j\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:45 crc kubenswrapper[4770]: I1209 12:52:45.658876 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:46 crc kubenswrapper[4770]: I1209 12:52:46.089177 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerStarted","Data":"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34"} Dec 09 12:52:46 crc kubenswrapper[4770]: I1209 12:52:46.122252 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.097262 4770 generic.go:334] "Generic (PLEG): container finished" podID="2dc25aed-199d-4603-839c-874eb52fbb37" containerID="4fcbed51e0d0b64dc08304f71af1d12ad867120d19519f5b4a2fea0dcba12f1d" exitCode=0 Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.097388 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerDied","Data":"4fcbed51e0d0b64dc08304f71af1d12ad867120d19519f5b4a2fea0dcba12f1d"} Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.098557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerStarted","Data":"fb625f98bd8f8e9212f303f9db1f2dba9805606d20d85489a56f5bb487bfdb70"} Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.101452 4770 generic.go:334] "Generic (PLEG): container finished" podID="239014c3-d863-42ab-8f3d-62f19934afba" containerID="3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec" exitCode=0 Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.101775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerDied","Data":"3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec"} Dec 09 12:52:47 crc kubenswrapper[4770]: I1209 12:52:47.120380 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2xm8j" podStartSLOduration=3.62384179 podStartE2EDuration="5.120339212s" podCreationTimestamp="2025-12-09 12:52:42 +0000 UTC" firstStartedPulling="2025-12-09 12:52:44.004230266 +0000 UTC m=+4889.244988785" lastFinishedPulling="2025-12-09 12:52:45.500727688 +0000 UTC m=+4890.741486207" observedRunningTime="2025-12-09 12:52:46.146334751 +0000 UTC m=+4891.387093270" watchObservedRunningTime="2025-12-09 12:52:47.120339212 +0000 UTC m=+4892.361097741" Dec 09 12:52:48 crc kubenswrapper[4770]: I1209 12:52:48.111358 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerStarted","Data":"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4"} Dec 09 12:52:48 crc kubenswrapper[4770]: I1209 12:52:48.113738 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerStarted","Data":"d59fea621a2d9bd0aab0a942046bba77396e41ac39c510620886def0adf88ac3"} Dec 09 12:52:48 crc kubenswrapper[4770]: I1209 12:52:48.131285 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xwkhb" podStartSLOduration=2.547153485 podStartE2EDuration="5.131260369s" podCreationTimestamp="2025-12-09 12:52:43 +0000 UTC" firstStartedPulling="2025-12-09 12:52:45.024584653 +0000 UTC m=+4890.265343172" lastFinishedPulling="2025-12-09 12:52:47.608691537 +0000 UTC m=+4892.849450056" observedRunningTime="2025-12-09 12:52:48.126582141 +0000 UTC m=+4893.367340660" watchObservedRunningTime="2025-12-09 12:52:48.131260369 +0000 UTC m=+4893.372018888" Dec 09 12:52:49 crc kubenswrapper[4770]: I1209 12:52:49.124079 4770 generic.go:334] "Generic (PLEG): container finished" podID="2dc25aed-199d-4603-839c-874eb52fbb37" containerID="d59fea621a2d9bd0aab0a942046bba77396e41ac39c510620886def0adf88ac3" exitCode=0 Dec 09 12:52:49 crc kubenswrapper[4770]: I1209 12:52:49.124189 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerDied","Data":"d59fea621a2d9bd0aab0a942046bba77396e41ac39c510620886def0adf88ac3"} Dec 09 12:52:50 crc kubenswrapper[4770]: I1209 12:52:50.144158 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerStarted","Data":"485e17f74fd57419fe6a02222d51aaa4cdc37319f96e5a47abf4e59febab7d53"} Dec 09 12:52:50 crc kubenswrapper[4770]: I1209 12:52:50.175377 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kpk4j" podStartSLOduration=2.7177263419999997 podStartE2EDuration="5.175354259s" podCreationTimestamp="2025-12-09 12:52:45 +0000 UTC" firstStartedPulling="2025-12-09 12:52:47.099181985 +0000 UTC m=+4892.339940504" lastFinishedPulling="2025-12-09 12:52:49.556809902 +0000 UTC m=+4894.797568421" observedRunningTime="2025-12-09 12:52:50.173297367 +0000 UTC m=+4895.414055896" watchObservedRunningTime="2025-12-09 12:52:50.175354259 +0000 UTC m=+4895.416112778" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.227745 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.228156 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.277427 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.421015 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.421072 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:53 crc kubenswrapper[4770]: I1209 12:52:53.473530 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:54 crc kubenswrapper[4770]: I1209 12:52:54.220207 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:54 crc kubenswrapper[4770]: I1209 12:52:54.239636 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:55 crc kubenswrapper[4770]: I1209 12:52:55.487226 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:55 crc kubenswrapper[4770]: I1209 12:52:55.659702 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:55 crc kubenswrapper[4770]: I1209 12:52:55.659789 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:55 crc kubenswrapper[4770]: I1209 12:52:55.703276 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:56 crc kubenswrapper[4770]: I1209 12:52:56.188451 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xwkhb" podUID="239014c3-d863-42ab-8f3d-62f19934afba" containerName="registry-server" containerID="cri-o://1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4" gracePeriod=2 Dec 09 12:52:56 crc kubenswrapper[4770]: I1209 12:52:56.239033 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:56 crc kubenswrapper[4770]: E1209 12:52:56.373546 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod239014c3_d863_42ab_8f3d_62f19934afba.slice/crio-1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4.scope\": RecentStats: unable to find data in memory cache]" Dec 09 12:52:56 crc kubenswrapper[4770]: I1209 12:52:56.485552 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:56 crc kubenswrapper[4770]: I1209 12:52:56.485787 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2xm8j" podUID="fc8c352a-3f30-4fc3-bb35-fe518cd4d940" containerName="registry-server" containerID="cri-o://b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34" gracePeriod=2 Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.720890 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.912832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-258ll\" (UniqueName: \"kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll\") pod \"239014c3-d863-42ab-8f3d-62f19934afba\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.912930 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content\") pod \"239014c3-d863-42ab-8f3d-62f19934afba\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.913040 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities\") pod \"239014c3-d863-42ab-8f3d-62f19934afba\" (UID: \"239014c3-d863-42ab-8f3d-62f19934afba\") " Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.917281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities" (OuterVolumeSpecName: "utilities") pod "239014c3-d863-42ab-8f3d-62f19934afba" (UID: "239014c3-d863-42ab-8f3d-62f19934afba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.923047 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll" (OuterVolumeSpecName: "kube-api-access-258ll") pod "239014c3-d863-42ab-8f3d-62f19934afba" (UID: "239014c3-d863-42ab-8f3d-62f19934afba"). InnerVolumeSpecName "kube-api-access-258ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:52:57 crc kubenswrapper[4770]: I1209 12:52:57.977313 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239014c3-d863-42ab-8f3d-62f19934afba" (UID: "239014c3-d863-42ab-8f3d-62f19934afba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.028847 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-258ll\" (UniqueName: \"kubernetes.io/projected/239014c3-d863-42ab-8f3d-62f19934afba-kube-api-access-258ll\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.028882 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.028912 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239014c3-d863-42ab-8f3d-62f19934afba-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.069656 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.129994 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndkx4\" (UniqueName: \"kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4\") pod \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.130145 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities\") pod \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.130198 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content\") pod \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\" (UID: \"fc8c352a-3f30-4fc3-bb35-fe518cd4d940\") " Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.143101 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities" (OuterVolumeSpecName: "utilities") pod "fc8c352a-3f30-4fc3-bb35-fe518cd4d940" (UID: "fc8c352a-3f30-4fc3-bb35-fe518cd4d940"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.145820 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4" (OuterVolumeSpecName: "kube-api-access-ndkx4") pod "fc8c352a-3f30-4fc3-bb35-fe518cd4d940" (UID: "fc8c352a-3f30-4fc3-bb35-fe518cd4d940"). InnerVolumeSpecName "kube-api-access-ndkx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.155808 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc8c352a-3f30-4fc3-bb35-fe518cd4d940" (UID: "fc8c352a-3f30-4fc3-bb35-fe518cd4d940"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.209254 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc8c352a-3f30-4fc3-bb35-fe518cd4d940" containerID="b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34" exitCode=0 Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.209303 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerDied","Data":"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34"} Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.209361 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xm8j" event={"ID":"fc8c352a-3f30-4fc3-bb35-fe518cd4d940","Type":"ContainerDied","Data":"aecddccaab60048b49f546c6fb79d05210bcb70428cc323b1b5b1569dffd3aec"} Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.209360 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xm8j" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.209395 4770 scope.go:117] "RemoveContainer" containerID="b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.213025 4770 generic.go:334] "Generic (PLEG): container finished" podID="239014c3-d863-42ab-8f3d-62f19934afba" containerID="1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4" exitCode=0 Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.213050 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerDied","Data":"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4"} Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.213068 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkhb" event={"ID":"239014c3-d863-42ab-8f3d-62f19934afba","Type":"ContainerDied","Data":"07a0f2af9fa97b27820ce78b0371b39268060c1a3d5ccd08ce4c9ec21db2a14e"} Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.213139 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkhb" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.241170 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.241214 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.241254 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndkx4\" (UniqueName: \"kubernetes.io/projected/fc8c352a-3f30-4fc3-bb35-fe518cd4d940-kube-api-access-ndkx4\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.242247 4770 scope.go:117] "RemoveContainer" containerID="76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.253165 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.273515 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xm8j"] Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.280273 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.282726 4770 scope.go:117] "RemoveContainer" containerID="fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.290204 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xwkhb"] Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.303619 4770 scope.go:117] "RemoveContainer" containerID="b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.304148 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34\": container with ID starting with b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34 not found: ID does not exist" containerID="b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.304208 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34"} err="failed to get container status \"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34\": rpc error: code = NotFound desc = could not find container \"b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34\": container with ID starting with b9410fe1ca49abcccd70bb935f899130f683fc0ba1db8b2b28311bbed618da34 not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.304308 4770 scope.go:117] "RemoveContainer" containerID="76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.304731 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8\": container with ID starting with 76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8 not found: ID does not exist" containerID="76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.304767 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8"} err="failed to get container status \"76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8\": rpc error: code = NotFound desc = could not find container \"76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8\": container with ID starting with 76bedb22ac4ac3fbd96981ef94b34b612bc27dc7528a1871a2a38094493880a8 not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.304798 4770 scope.go:117] "RemoveContainer" containerID="fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.305319 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca\": container with ID starting with fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca not found: ID does not exist" containerID="fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.305352 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca"} err="failed to get container status \"fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca\": rpc error: code = NotFound desc = could not find container \"fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca\": container with ID starting with fc2e5faa676fc4669a884adfb17fc619202d3cbbff6c8288329ae61b58bc11ca not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.305377 4770 scope.go:117] "RemoveContainer" containerID="1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.320759 4770 scope.go:117] "RemoveContainer" containerID="3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.344804 4770 scope.go:117] "RemoveContainer" containerID="a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.363065 4770 scope.go:117] "RemoveContainer" containerID="1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.363571 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4\": container with ID starting with 1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4 not found: ID does not exist" containerID="1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.363601 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4"} err="failed to get container status \"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4\": rpc error: code = NotFound desc = could not find container \"1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4\": container with ID starting with 1be7a631a2c8d82b026888aa2fb57725778ba622f805172022a9ecc5331025f4 not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.363622 4770 scope.go:117] "RemoveContainer" containerID="3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.363891 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec\": container with ID starting with 3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec not found: ID does not exist" containerID="3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.363966 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec"} err="failed to get container status \"3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec\": rpc error: code = NotFound desc = could not find container \"3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec\": container with ID starting with 3c07cb199bb6573ad37d9363453fc9d60f502ede5f2f7824b8489b2d43c086ec not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.364017 4770 scope.go:117] "RemoveContainer" containerID="a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c" Dec 09 12:52:58 crc kubenswrapper[4770]: E1209 12:52:58.364628 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c\": container with ID starting with a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c not found: ID does not exist" containerID="a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.364654 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c"} err="failed to get container status \"a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c\": rpc error: code = NotFound desc = could not find container \"a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c\": container with ID starting with a558522d0b6b62ebfd0045fa5ebefe6a957a27d3ace096d52c858d59e9f1e81c not found: ID does not exist" Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.889130 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:52:58 crc kubenswrapper[4770]: I1209 12:52:58.889698 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kpk4j" podUID="2dc25aed-199d-4603-839c-874eb52fbb37" containerName="registry-server" containerID="cri-o://485e17f74fd57419fe6a02222d51aaa4cdc37319f96e5a47abf4e59febab7d53" gracePeriod=2 Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.222355 4770 generic.go:334] "Generic (PLEG): container finished" podID="2dc25aed-199d-4603-839c-874eb52fbb37" containerID="485e17f74fd57419fe6a02222d51aaa4cdc37319f96e5a47abf4e59febab7d53" exitCode=0 Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.222411 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerDied","Data":"485e17f74fd57419fe6a02222d51aaa4cdc37319f96e5a47abf4e59febab7d53"} Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.350958 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239014c3-d863-42ab-8f3d-62f19934afba" path="/var/lib/kubelet/pods/239014c3-d863-42ab-8f3d-62f19934afba/volumes" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.351708 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8c352a-3f30-4fc3-bb35-fe518cd4d940" path="/var/lib/kubelet/pods/fc8c352a-3f30-4fc3-bb35-fe518cd4d940/volumes" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.773693 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.890586 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kp98\" (UniqueName: \"kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98\") pod \"2dc25aed-199d-4603-839c-874eb52fbb37\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.890671 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities\") pod \"2dc25aed-199d-4603-839c-874eb52fbb37\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.890702 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content\") pod \"2dc25aed-199d-4603-839c-874eb52fbb37\" (UID: \"2dc25aed-199d-4603-839c-874eb52fbb37\") " Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.892950 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities" (OuterVolumeSpecName: "utilities") pod "2dc25aed-199d-4603-839c-874eb52fbb37" (UID: "2dc25aed-199d-4603-839c-874eb52fbb37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.900117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98" (OuterVolumeSpecName: "kube-api-access-2kp98") pod "2dc25aed-199d-4603-839c-874eb52fbb37" (UID: "2dc25aed-199d-4603-839c-874eb52fbb37"). InnerVolumeSpecName "kube-api-access-2kp98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.941135 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dc25aed-199d-4603-839c-874eb52fbb37" (UID: "2dc25aed-199d-4603-839c-874eb52fbb37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.992492 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kp98\" (UniqueName: \"kubernetes.io/projected/2dc25aed-199d-4603-839c-874eb52fbb37-kube-api-access-2kp98\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.992538 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 12:52:59 crc kubenswrapper[4770]: I1209 12:52:59.992555 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc25aed-199d-4603-839c-874eb52fbb37-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.236821 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kpk4j" event={"ID":"2dc25aed-199d-4603-839c-874eb52fbb37","Type":"ContainerDied","Data":"fb625f98bd8f8e9212f303f9db1f2dba9805606d20d85489a56f5bb487bfdb70"} Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.236893 4770 scope.go:117] "RemoveContainer" containerID="485e17f74fd57419fe6a02222d51aaa4cdc37319f96e5a47abf4e59febab7d53" Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.236965 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kpk4j" Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.370014 4770 scope.go:117] "RemoveContainer" containerID="d59fea621a2d9bd0aab0a942046bba77396e41ac39c510620886def0adf88ac3" Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.374637 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.382697 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kpk4j"] Dec 09 12:53:00 crc kubenswrapper[4770]: I1209 12:53:00.393731 4770 scope.go:117] "RemoveContainer" containerID="4fcbed51e0d0b64dc08304f71af1d12ad867120d19519f5b4a2fea0dcba12f1d" Dec 09 12:53:01 crc kubenswrapper[4770]: I1209 12:53:01.350107 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc25aed-199d-4603-839c-874eb52fbb37" path="/var/lib/kubelet/pods/2dc25aed-199d-4603-839c-874eb52fbb37/volumes" Dec 09 12:53:02 crc kubenswrapper[4770]: I1209 12:53:02.473270 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:53:02 crc kubenswrapper[4770]: I1209 12:53:02.473632 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.474349 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.474973 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.475045 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.475961 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.476050 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6" gracePeriod=600 Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.703251 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6" exitCode=0 Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.703299 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6"} Dec 09 12:53:32 crc kubenswrapper[4770]: I1209 12:53:32.703655 4770 scope.go:117] "RemoveContainer" containerID="b6bef1b5f449e61fa8bd671d50270483dc4f34a767f6ba21dc32f604515c644f" Dec 09 12:53:33 crc kubenswrapper[4770]: I1209 12:53:33.713050 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerStarted","Data":"a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2"} Dec 09 12:53:54 crc kubenswrapper[4770]: I1209 12:53:54.925321 4770 generic.go:334] "Generic (PLEG): container finished" podID="c1467a75-7b8e-4b1d-a56e-9529a26a1c08" containerID="17a9922e58d554d52e910815c39297ca4ee464ebd94048d073bc85a3c132abb3" exitCode=0 Dec 09 12:53:54 crc kubenswrapper[4770]: I1209 12:53:54.925405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" event={"ID":"c1467a75-7b8e-4b1d-a56e-9529a26a1c08","Type":"ContainerDied","Data":"17a9922e58d554d52e910815c39297ca4ee464ebd94048d073bc85a3c132abb3"} Dec 09 12:53:54 crc kubenswrapper[4770]: I1209 12:53:54.926751 4770 scope.go:117] "RemoveContainer" containerID="17a9922e58d554d52e910815c39297ca4ee464ebd94048d073bc85a3c132abb3" Dec 09 12:53:55 crc kubenswrapper[4770]: I1209 12:53:55.584109 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pcrp_must-gather-cxm8p_c1467a75-7b8e-4b1d-a56e-9529a26a1c08/gather/0.log" Dec 09 12:54:03 crc kubenswrapper[4770]: I1209 12:54:03.769654 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2pcrp/must-gather-cxm8p"] Dec 09 12:54:03 crc kubenswrapper[4770]: I1209 12:54:03.770713 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" podUID="c1467a75-7b8e-4b1d-a56e-9529a26a1c08" containerName="copy" containerID="cri-o://fd2abc20aca6fe97ef4c9ec3962f49c4c5ab1afc25959ecdb9a7ee4f114ab9b6" gracePeriod=2 Dec 09 12:54:03 crc kubenswrapper[4770]: I1209 12:54:03.779214 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2pcrp/must-gather-cxm8p"] Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.006766 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pcrp_must-gather-cxm8p_c1467a75-7b8e-4b1d-a56e-9529a26a1c08/copy/0.log" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.007415 4770 generic.go:334] "Generic (PLEG): container finished" podID="c1467a75-7b8e-4b1d-a56e-9529a26a1c08" containerID="fd2abc20aca6fe97ef4c9ec3962f49c4c5ab1afc25959ecdb9a7ee4f114ab9b6" exitCode=143 Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.455872 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pcrp_must-gather-cxm8p_c1467a75-7b8e-4b1d-a56e-9529a26a1c08/copy/0.log" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.456492 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.584174 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output\") pod \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.584220 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcfkk\" (UniqueName: \"kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk\") pod \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\" (UID: \"c1467a75-7b8e-4b1d-a56e-9529a26a1c08\") " Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.588994 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk" (OuterVolumeSpecName: "kube-api-access-zcfkk") pod "c1467a75-7b8e-4b1d-a56e-9529a26a1c08" (UID: "c1467a75-7b8e-4b1d-a56e-9529a26a1c08"). InnerVolumeSpecName "kube-api-access-zcfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.676403 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c1467a75-7b8e-4b1d-a56e-9529a26a1c08" (UID: "c1467a75-7b8e-4b1d-a56e-9529a26a1c08"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.686321 4770 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 09 12:54:04 crc kubenswrapper[4770]: I1209 12:54:04.686349 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcfkk\" (UniqueName: \"kubernetes.io/projected/c1467a75-7b8e-4b1d-a56e-9529a26a1c08-kube-api-access-zcfkk\") on node \"crc\" DevicePath \"\"" Dec 09 12:54:05 crc kubenswrapper[4770]: I1209 12:54:05.200638 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2pcrp_must-gather-cxm8p_c1467a75-7b8e-4b1d-a56e-9529a26a1c08/copy/0.log" Dec 09 12:54:05 crc kubenswrapper[4770]: I1209 12:54:05.201328 4770 scope.go:117] "RemoveContainer" containerID="fd2abc20aca6fe97ef4c9ec3962f49c4c5ab1afc25959ecdb9a7ee4f114ab9b6" Dec 09 12:54:05 crc kubenswrapper[4770]: I1209 12:54:05.201488 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2pcrp/must-gather-cxm8p" Dec 09 12:54:05 crc kubenswrapper[4770]: I1209 12:54:05.222931 4770 scope.go:117] "RemoveContainer" containerID="17a9922e58d554d52e910815c39297ca4ee464ebd94048d073bc85a3c132abb3" Dec 09 12:54:05 crc kubenswrapper[4770]: I1209 12:54:05.351060 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1467a75-7b8e-4b1d-a56e-9529a26a1c08" path="/var/lib/kubelet/pods/c1467a75-7b8e-4b1d-a56e-9529a26a1c08/volumes" Dec 09 12:55:32 crc kubenswrapper[4770]: I1209 12:55:32.473612 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:55:32 crc kubenswrapper[4770]: I1209 12:55:32.474487 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:56:02 crc kubenswrapper[4770]: I1209 12:56:02.474047 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:56:02 crc kubenswrapper[4770]: I1209 12:56:02.474709 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:56:31 crc kubenswrapper[4770]: I1209 12:56:31.938894 4770 scope.go:117] "RemoveContainer" containerID="4263ac399530190891244e9dd73ee6d2b37cb0d0238032c1a6ff33a4421646b9" Dec 09 12:56:31 crc kubenswrapper[4770]: I1209 12:56:31.978504 4770 scope.go:117] "RemoveContainer" containerID="88031c6defa364721c70c5e92c7d32636a9b24a41cc24c03860cd5e1dc780bad" Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.006551 4770 scope.go:117] "RemoveContainer" containerID="350e709db9ee32fbb6a5dabcc17ed5a2d9aed3e20d3df253cdf872c879aee0c3" Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.473595 4770 patch_prober.go:28] interesting pod/machine-config-daemon-qxvgc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.474192 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.474258 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.475173 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2"} pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 12:56:32 crc kubenswrapper[4770]: I1209 12:56:32.475245 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerName="machine-config-daemon" containerID="cri-o://a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" gracePeriod=600 Dec 09 12:56:32 crc kubenswrapper[4770]: E1209 12:56:32.750775 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:56:33 crc kubenswrapper[4770]: I1209 12:56:33.009131 4770 generic.go:334] "Generic (PLEG): container finished" podID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" exitCode=0 Dec 09 12:56:33 crc kubenswrapper[4770]: I1209 12:56:33.009210 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" event={"ID":"18e4a331-a41b-419a-9a35-f9fb57ff38f1","Type":"ContainerDied","Data":"a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2"} Dec 09 12:56:33 crc kubenswrapper[4770]: I1209 12:56:33.010177 4770 scope.go:117] "RemoveContainer" containerID="17cedc8fe96d20775c4cb60dbdb021c8cb4e970ce685fc3c618c7af4c2989da6" Dec 09 12:56:33 crc kubenswrapper[4770]: I1209 12:56:33.010980 4770 scope.go:117] "RemoveContainer" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" Dec 09 12:56:33 crc kubenswrapper[4770]: E1209 12:56:33.011449 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:56:44 crc kubenswrapper[4770]: I1209 12:56:44.339973 4770 scope.go:117] "RemoveContainer" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" Dec 09 12:56:44 crc kubenswrapper[4770]: E1209 12:56:44.340806 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:56:55 crc kubenswrapper[4770]: I1209 12:56:55.344466 4770 scope.go:117] "RemoveContainer" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" Dec 09 12:56:55 crc kubenswrapper[4770]: E1209 12:56:55.345229 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:57:07 crc kubenswrapper[4770]: I1209 12:57:07.471499 4770 scope.go:117] "RemoveContainer" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" Dec 09 12:57:07 crc kubenswrapper[4770]: E1209 12:57:07.472209 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" Dec 09 12:57:19 crc kubenswrapper[4770]: I1209 12:57:19.375766 4770 scope.go:117] "RemoveContainer" containerID="a4e0b0071b1582aab8788f310abc36df5c4de569cbbd32d12ae272b9a43684b2" Dec 09 12:57:19 crc kubenswrapper[4770]: E1209 12:57:19.376587 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qxvgc_openshift-machine-config-operator(18e4a331-a41b-419a-9a35-f9fb57ff38f1)\"" pod="openshift-machine-config-operator/machine-config-daemon-qxvgc" podUID="18e4a331-a41b-419a-9a35-f9fb57ff38f1" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116016301024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116016302017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116003660016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116003661015454 5ustar corecore